var/home/core/zuul-output/0000755000175000017500000000000015071202710014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071207131015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004031467115071207121017677 0ustar rootrootOct 07 12:24:02 crc systemd[1]: Starting Kubernetes Kubelet... Oct 07 12:24:02 crc restorecon[4672]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:02 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 12:24:03 crc restorecon[4672]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 07 12:24:04 crc kubenswrapper[4702]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 12:24:04 crc kubenswrapper[4702]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 07 12:24:04 crc kubenswrapper[4702]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 12:24:04 crc kubenswrapper[4702]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 12:24:04 crc kubenswrapper[4702]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 07 12:24:04 crc kubenswrapper[4702]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.316647 4702 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326651 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326703 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326713 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326722 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326732 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326743 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326756 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326768 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326777 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326786 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326794 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326803 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326812 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326820 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326829 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326838 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326846 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326860 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326869 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326878 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326887 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326919 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326927 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326936 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326947 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326959 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326968 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326977 4702 feature_gate.go:330] unrecognized feature gate: Example Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326986 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.326994 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327005 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327016 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327025 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327034 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327042 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327050 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327059 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327067 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327075 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327085 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327093 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327102 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327114 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327123 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327133 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327142 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327151 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327161 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327170 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327178 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327186 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327194 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327203 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327211 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327219 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327228 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327237 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327246 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327254 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327262 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327273 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327281 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327291 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327299 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327307 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327315 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327324 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327332 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327340 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327348 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.327356 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329214 4702 flags.go:64] FLAG: --address="0.0.0.0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329241 4702 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329258 4702 flags.go:64] FLAG: --anonymous-auth="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329270 4702 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329283 4702 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329293 4702 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329306 4702 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329325 4702 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329336 4702 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329346 4702 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329356 4702 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329369 4702 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329380 4702 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329390 4702 flags.go:64] FLAG: --cgroup-root="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329400 4702 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329410 4702 flags.go:64] FLAG: --client-ca-file="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329420 4702 flags.go:64] FLAG: --cloud-config="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329431 4702 flags.go:64] FLAG: --cloud-provider="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329440 4702 flags.go:64] FLAG: --cluster-dns="[]" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329455 4702 flags.go:64] FLAG: --cluster-domain="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329465 4702 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329475 4702 flags.go:64] FLAG: --config-dir="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329485 4702 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329496 4702 flags.go:64] FLAG: --container-log-max-files="5" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329508 4702 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329518 4702 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329528 4702 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329538 4702 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329551 4702 flags.go:64] FLAG: --contention-profiling="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329561 4702 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329570 4702 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329580 4702 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329592 4702 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329605 4702 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329617 4702 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329627 4702 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329637 4702 flags.go:64] FLAG: --enable-load-reader="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329647 4702 flags.go:64] FLAG: --enable-server="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329657 4702 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329670 4702 flags.go:64] FLAG: --event-burst="100" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329680 4702 flags.go:64] FLAG: --event-qps="50" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329689 4702 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329698 4702 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329708 4702 flags.go:64] FLAG: --eviction-hard="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329720 4702 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329730 4702 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329739 4702 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329751 4702 flags.go:64] FLAG: --eviction-soft="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329761 4702 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329771 4702 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329781 4702 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329790 4702 flags.go:64] FLAG: --experimental-mounter-path="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329799 4702 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329809 4702 flags.go:64] FLAG: --fail-swap-on="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329819 4702 flags.go:64] FLAG: --feature-gates="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329831 4702 flags.go:64] FLAG: --file-check-frequency="20s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329841 4702 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329851 4702 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329861 4702 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329871 4702 flags.go:64] FLAG: --healthz-port="10248" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329881 4702 flags.go:64] FLAG: --help="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.329891 4702 flags.go:64] FLAG: --hostname-override="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330062 4702 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330095 4702 flags.go:64] FLAG: --http-check-frequency="20s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330115 4702 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330127 4702 flags.go:64] FLAG: --image-credential-provider-config="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330135 4702 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330147 4702 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330154 4702 flags.go:64] FLAG: --image-service-endpoint="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330162 4702 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330172 4702 flags.go:64] FLAG: --kube-api-burst="100" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330184 4702 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330190 4702 flags.go:64] FLAG: --kube-api-qps="50" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330198 4702 flags.go:64] FLAG: --kube-reserved="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330207 4702 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330215 4702 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330223 4702 flags.go:64] FLAG: --kubelet-cgroups="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330230 4702 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330236 4702 flags.go:64] FLAG: --lock-file="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330242 4702 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330249 4702 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330255 4702 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330268 4702 flags.go:64] FLAG: --log-json-split-stream="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330278 4702 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330284 4702 flags.go:64] FLAG: --log-text-split-stream="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330291 4702 flags.go:64] FLAG: --logging-format="text" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330297 4702 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330304 4702 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330310 4702 flags.go:64] FLAG: --manifest-url="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330316 4702 flags.go:64] FLAG: --manifest-url-header="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330363 4702 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330372 4702 flags.go:64] FLAG: --max-open-files="1000000" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330382 4702 flags.go:64] FLAG: --max-pods="110" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330391 4702 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330398 4702 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330406 4702 flags.go:64] FLAG: --memory-manager-policy="None" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330414 4702 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330422 4702 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330430 4702 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330439 4702 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330464 4702 flags.go:64] FLAG: --node-status-max-images="50" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330472 4702 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330479 4702 flags.go:64] FLAG: --oom-score-adj="-999" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330487 4702 flags.go:64] FLAG: --pod-cidr="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330494 4702 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330506 4702 flags.go:64] FLAG: --pod-manifest-path="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330513 4702 flags.go:64] FLAG: --pod-max-pids="-1" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330520 4702 flags.go:64] FLAG: --pods-per-core="0" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330526 4702 flags.go:64] FLAG: --port="10250" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330533 4702 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330540 4702 flags.go:64] FLAG: --provider-id="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330547 4702 flags.go:64] FLAG: --qos-reserved="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330553 4702 flags.go:64] FLAG: --read-only-port="10255" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330560 4702 flags.go:64] FLAG: --register-node="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330567 4702 flags.go:64] FLAG: --register-schedulable="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330572 4702 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330592 4702 flags.go:64] FLAG: --registry-burst="10" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330599 4702 flags.go:64] FLAG: --registry-qps="5" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330606 4702 flags.go:64] FLAG: --reserved-cpus="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330619 4702 flags.go:64] FLAG: --reserved-memory="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330629 4702 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330636 4702 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330643 4702 flags.go:64] FLAG: --rotate-certificates="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330649 4702 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330655 4702 flags.go:64] FLAG: --runonce="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330661 4702 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330667 4702 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330674 4702 flags.go:64] FLAG: --seccomp-default="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330679 4702 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330685 4702 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330695 4702 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330705 4702 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330716 4702 flags.go:64] FLAG: --storage-driver-password="root" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330728 4702 flags.go:64] FLAG: --storage-driver-secure="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330738 4702 flags.go:64] FLAG: --storage-driver-table="stats" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330748 4702 flags.go:64] FLAG: --storage-driver-user="root" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330758 4702 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330769 4702 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330779 4702 flags.go:64] FLAG: --system-cgroups="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330788 4702 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330804 4702 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330814 4702 flags.go:64] FLAG: --tls-cert-file="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330823 4702 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330843 4702 flags.go:64] FLAG: --tls-min-version="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330853 4702 flags.go:64] FLAG: --tls-private-key-file="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330863 4702 flags.go:64] FLAG: --topology-manager-policy="none" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330872 4702 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330883 4702 flags.go:64] FLAG: --topology-manager-scope="container" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330893 4702 flags.go:64] FLAG: --v="2" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330952 4702 flags.go:64] FLAG: --version="false" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330966 4702 flags.go:64] FLAG: --vmodule="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330978 4702 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.330989 4702 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331277 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331290 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331300 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331310 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331319 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331328 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331337 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331346 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331354 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331366 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331378 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331388 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331397 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331407 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331417 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331427 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331437 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331446 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331454 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331463 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331473 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331482 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331494 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331504 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331512 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331521 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331530 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331538 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331546 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331555 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331563 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331574 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331585 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331596 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331605 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331615 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331626 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331638 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331650 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331659 4702 feature_gate.go:330] unrecognized feature gate: Example Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331668 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331677 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331686 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331694 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331702 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331710 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331719 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331727 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331750 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331759 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331767 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331776 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331784 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331792 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331801 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331809 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331818 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331826 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331835 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331843 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331852 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331860 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331868 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331876 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331885 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331926 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331936 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331953 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331962 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331970 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.331978 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.332007 4702 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.345610 4702 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.345675 4702 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345838 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345868 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345882 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345922 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345932 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345941 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345951 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345959 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345968 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345977 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345985 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.345994 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346002 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346011 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346020 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346028 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346036 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346045 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346055 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346067 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346077 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346087 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346097 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346105 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346114 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346123 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346131 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346140 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346148 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346160 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346168 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346177 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346186 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346195 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346204 4702 feature_gate.go:330] unrecognized feature gate: Example Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346212 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346221 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346229 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346238 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346247 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346255 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346266 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346276 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346286 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346296 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346306 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346349 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346359 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346370 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346382 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346393 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346401 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346414 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346425 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346435 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346444 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346452 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346461 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346469 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346478 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346487 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346495 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346503 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346512 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346521 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346531 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346540 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346548 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346556 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346565 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346573 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.346590 4702 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346835 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346847 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346858 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346868 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346877 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346887 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346919 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346930 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346940 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346949 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346959 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346968 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346976 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346985 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.346993 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347001 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347010 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347018 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347026 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347037 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347048 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347058 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347068 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347078 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347088 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347097 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347105 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347114 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347122 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347131 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347140 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347149 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347159 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347167 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347176 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347185 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347193 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347201 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347209 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347218 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347226 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347235 4702 feature_gate.go:330] unrecognized feature gate: Example Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347244 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347253 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347262 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347271 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347282 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347293 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347302 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347311 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347321 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347331 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347340 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347348 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347357 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347366 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347374 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347383 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347391 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347399 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347408 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347417 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347426 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347434 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347443 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347452 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347461 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347470 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347481 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347492 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.347502 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.347517 4702 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.347811 4702 server.go:940] "Client rotation is on, will bootstrap in background" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.353884 4702 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.354057 4702 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.356156 4702 server.go:997] "Starting client certificate rotation" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.356208 4702 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.356414 4702 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-15 02:18:18.905841512 +0000 UTC Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.356557 4702 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1645h54m14.549286694s for next certificate rotation Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.394386 4702 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.397763 4702 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.426038 4702 log.go:25] "Validated CRI v1 runtime API" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.470388 4702 log.go:25] "Validated CRI v1 image API" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.473282 4702 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.482411 4702 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-07-12-20-29-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.482463 4702 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.501471 4702 manager.go:217] Machine: {Timestamp:2025-10-07 12:24:04.49873256 +0000 UTC m=+0.674468686 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:bbdbab7f-ca5e-4664-b776-056ee45e0f4d BootID:765d68a7-0b5e-4d6a-980f-84924188d680 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ce:43:87 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ce:43:87 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:09:74:df Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:98:27:b3 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:83:be:5b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ec:82:6b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ca:0f:c0:0b:a1:57 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:c2:74:ab:b8:b5:d4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.501709 4702 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.501882 4702 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.502181 4702 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.502345 4702 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.502373 4702 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.502576 4702 topology_manager.go:138] "Creating topology manager with none policy" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.502587 4702 container_manager_linux.go:303] "Creating device plugin manager" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.503284 4702 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.503316 4702 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.504214 4702 state_mem.go:36] "Initialized new in-memory state store" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.504306 4702 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.508690 4702 kubelet.go:418] "Attempting to sync node with API server" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.508714 4702 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.508729 4702 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.508743 4702 kubelet.go:324] "Adding apiserver pod source" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.508756 4702 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.513560 4702 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.514697 4702 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.517645 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.517700 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.517863 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.517876 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.518058 4702 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520549 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520595 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520611 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520628 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520658 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520677 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520696 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520720 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520767 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520782 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520836 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.520851 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.522044 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.523649 4702 server.go:1280] "Started kubelet" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.524574 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.525370 4702 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.525394 4702 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 07 12:24:04 crc systemd[1]: Started Kubernetes Kubelet. Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.527850 4702 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.529672 4702 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.529856 4702 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.529937 4702 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 01:18:25.686425843 +0000 UTC Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.530072 4702 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2292h54m21.156358973s for next certificate rotation Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.538992 4702 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.539136 4702 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.539152 4702 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.539425 4702 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.540724 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.540872 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.529681 4702 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.190:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c3503ba5f1422 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 12:24:04.522988578 +0000 UTC m=+0.698724754,LastTimestamp:2025-10-07 12:24:04.522988578 +0000 UTC m=+0.698724754,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.543847 4702 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.543881 4702 factory.go:55] Registering systemd factory Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.543890 4702 factory.go:221] Registration of the systemd container factory successfully Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.544566 4702 factory.go:153] Registering CRI-O factory Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.544664 4702 factory.go:221] Registration of the crio container factory successfully Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.544725 4702 factory.go:103] Registering Raw factory Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.544865 4702 manager.go:1196] Started watching for new ooms in manager Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.574306 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.190:6443: connect: connection refused" interval="200ms" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.574506 4702 manager.go:319] Starting recovery of all containers Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.574633 4702 server.go:460] "Adding debug handlers to kubelet server" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579764 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579827 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579839 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579849 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579857 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579866 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579877 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579886 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579911 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579933 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579942 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579952 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579967 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579977 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579987 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.579998 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580007 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580017 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580053 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580064 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580074 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580085 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580096 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580107 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580119 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580130 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580142 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580154 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580164 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580173 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580184 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580194 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580203 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580214 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580227 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580236 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580247 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580257 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580268 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580281 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580298 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580309 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580319 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580328 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580339 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580349 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580361 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580371 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580381 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580391 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580402 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580414 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580429 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580441 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580451 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580462 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580474 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580485 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580496 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580507 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580516 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580527 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580538 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580549 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580560 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580571 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580582 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580592 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580602 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580612 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580622 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580632 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580643 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580655 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580665 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580677 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580687 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580699 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580711 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580722 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580732 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580743 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580754 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580764 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580774 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580785 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580794 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580806 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580816 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580827 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580837 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580847 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580856 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580866 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580876 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580886 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580909 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580919 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580928 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580938 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580948 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580959 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580969 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580979 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.580998 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581010 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581021 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581033 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581044 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581056 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581069 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581080 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581090 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581102 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581113 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581123 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581133 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581143 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581183 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581194 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581204 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581216 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581227 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581236 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581247 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581257 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581267 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581277 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581289 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581299 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581309 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581319 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581331 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581341 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581351 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581362 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581375 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581385 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581395 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581405 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581415 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581424 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581435 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581446 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581456 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581467 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581477 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581488 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581498 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581507 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581515 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581525 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581533 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581544 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581553 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581561 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581570 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581580 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581590 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581600 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581608 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581618 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581627 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581636 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581646 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581655 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581663 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581672 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581684 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581693 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581701 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581710 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581719 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581729 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581739 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581749 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581758 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581767 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581776 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581786 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.581795 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589601 4702 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589639 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589655 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589671 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589686 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589699 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589711 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589722 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589732 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589745 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589762 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589776 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589790 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589802 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589813 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589825 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589835 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589845 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589854 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589867 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589878 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589887 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589915 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589927 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589937 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589947 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589957 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589966 4702 reconstruct.go:97] "Volume reconstruction finished" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.589973 4702 reconciler.go:26] "Reconciler: start to sync state" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.608959 4702 manager.go:324] Recovery completed Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.623688 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.625795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.626252 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.626279 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.627447 4702 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.627482 4702 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.627539 4702 state_mem.go:36] "Initialized new in-memory state store" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.635668 4702 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.639353 4702 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.639415 4702 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.639488 4702 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.639544 4702 kubelet.go:2335] "Starting kubelet main sync loop" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.639637 4702 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 07 12:24:04 crc kubenswrapper[4702]: W1007 12:24:04.640554 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.640716 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.699439 4702 policy_none.go:49] "None policy: Start" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.701610 4702 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.701677 4702 state_mem.go:35] "Initializing new in-memory state store" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.740046 4702 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.740228 4702 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.776090 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.190:6443: connect: connection refused" interval="400ms" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.788661 4702 manager.go:334] "Starting Device Plugin manager" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.788748 4702 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.788772 4702 server.go:79] "Starting device plugin registration server" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.789539 4702 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.789571 4702 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.789758 4702 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.789950 4702 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.789967 4702 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.801478 4702 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.890986 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.892761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.892836 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.892853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.892949 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:04 crc kubenswrapper[4702]: E1007 12:24:04.893783 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.190:6443: connect: connection refused" node="crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.940429 4702 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.940567 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.942136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.942194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.942211 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.942398 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.942788 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.942924 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.943501 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.943553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.943568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.943755 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.944096 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.944162 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.945029 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.945083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.945110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.946239 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.946289 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.946307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.946525 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.946623 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.946655 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.947848 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.947872 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.947883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948367 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948416 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948637 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948758 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.948808 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.949764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.949796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.949812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.950133 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.950177 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.949778 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.950590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.950614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.951064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.951086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.951099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.995294 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:04 crc kubenswrapper[4702]: I1007 12:24:04.995353 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.094855 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.096611 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.096804 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.096849 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097034 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097141 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097223 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097301 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097386 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097463 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097550 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097634 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097713 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.097811 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.098017 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.098126 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.098209 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.099261 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.102153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.102221 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.102243 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.102291 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.102975 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.190:6443: connect: connection refused" node="crc" Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.177073 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.190:6443: connect: connection refused" interval="800ms" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200538 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200639 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200711 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200759 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200788 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200826 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200943 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200961 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200960 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201010 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200959 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200856 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.200975 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201109 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201164 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201053 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201193 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201219 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201229 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201249 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201272 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201276 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201314 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201344 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.201390 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.276542 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.284480 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.301978 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.324077 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.326976 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-990a5e211bbbd8e12e3c5ec6d1900c442b5c5ecafe9a9c10c8b8517618fcc8b6 WatchSource:0}: Error finding container 990a5e211bbbd8e12e3c5ec6d1900c442b5c5ecafe9a9c10c8b8517618fcc8b6: Status 404 returned error can't find the container with id 990a5e211bbbd8e12e3c5ec6d1900c442b5c5ecafe9a9c10c8b8517618fcc8b6 Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.328237 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-cdf00b4ab91e501255d8b8a0c96f49f92325cb2884da7ffa7630e31da0c739a6 WatchSource:0}: Error finding container cdf00b4ab91e501255d8b8a0c96f49f92325cb2884da7ffa7630e31da0c739a6: Status 404 returned error can't find the container with id cdf00b4ab91e501255d8b8a0c96f49f92325cb2884da7ffa7630e31da0c739a6 Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.328701 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.329193 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-25a83b14a4ff9cb8484aa4815d057ae8f76f41d27397efd0b4be4aa184b1f12f WatchSource:0}: Error finding container 25a83b14a4ff9cb8484aa4815d057ae8f76f41d27397efd0b4be4aa184b1f12f: Status 404 returned error can't find the container with id 25a83b14a4ff9cb8484aa4815d057ae8f76f41d27397efd0b4be4aa184b1f12f Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.342414 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-2c5ac0315aa8e9db8ac15ecddf5535fb7894c04995adb1e09a3ffddb906e17db WatchSource:0}: Error finding container 2c5ac0315aa8e9db8ac15ecddf5535fb7894c04995adb1e09a3ffddb906e17db: Status 404 returned error can't find the container with id 2c5ac0315aa8e9db8ac15ecddf5535fb7894c04995adb1e09a3ffddb906e17db Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.359578 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-22fef11a3f81b422e2d1c954fc8ea29ad98cd301f08bdc1378819d1db0da859e WatchSource:0}: Error finding container 22fef11a3f81b422e2d1c954fc8ea29ad98cd301f08bdc1378819d1db0da859e: Status 404 returned error can't find the container with id 22fef11a3f81b422e2d1c954fc8ea29ad98cd301f08bdc1378819d1db0da859e Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.365711 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.365801 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.503367 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.505099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.505155 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.505172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.505209 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.505786 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.190:6443: connect: connection refused" node="crc" Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.525732 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.645596 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"990a5e211bbbd8e12e3c5ec6d1900c442b5c5ecafe9a9c10c8b8517618fcc8b6"} Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.647549 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"22fef11a3f81b422e2d1c954fc8ea29ad98cd301f08bdc1378819d1db0da859e"} Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.649059 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2c5ac0315aa8e9db8ac15ecddf5535fb7894c04995adb1e09a3ffddb906e17db"} Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.651478 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"25a83b14a4ff9cb8484aa4815d057ae8f76f41d27397efd0b4be4aa184b1f12f"} Oct 07 12:24:05 crc kubenswrapper[4702]: I1007 12:24:05.655168 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cdf00b4ab91e501255d8b8a0c96f49f92325cb2884da7ffa7630e31da0c739a6"} Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.748654 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.748800 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:05 crc kubenswrapper[4702]: W1007 12:24:05.828812 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.829029 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:05 crc kubenswrapper[4702]: E1007 12:24:05.979661 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.190:6443: connect: connection refused" interval="1.6s" Oct 07 12:24:06 crc kubenswrapper[4702]: W1007 12:24:06.045530 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:06 crc kubenswrapper[4702]: E1007 12:24:06.045688 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.307000 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.309521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.309586 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.309609 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.309660 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:06 crc kubenswrapper[4702]: E1007 12:24:06.310407 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.190:6443: connect: connection refused" node="crc" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.525440 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.659823 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12" exitCode=0 Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.659957 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.659966 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12"} Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.661011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.661054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.661067 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.662792 4702 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9b86cf396b84e1a8a67c39d95f5c06c283ef3c873bbd8384e8cddef2b770fdb2" exitCode=0 Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.662879 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9b86cf396b84e1a8a67c39d95f5c06c283ef3c873bbd8384e8cddef2b770fdb2"} Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.663622 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.666559 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.667984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.668042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.668060 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.668818 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.668836 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.668845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.669822 4702 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402" exitCode=0 Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.669880 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402"} Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.669972 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.671029 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.671073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.671088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.671654 4702 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c" exitCode=0 Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.671762 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.671728 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c"} Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.673106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.673149 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.673165 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.677306 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8"} Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.677331 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258"} Oct 07 12:24:06 crc kubenswrapper[4702]: I1007 12:24:06.677341 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585"} Oct 07 12:24:07 crc kubenswrapper[4702]: E1007 12:24:07.238341 4702 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.190:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c3503ba5f1422 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 12:24:04.522988578 +0000 UTC m=+0.698724754,LastTimestamp:2025-10-07 12:24:04.522988578 +0000 UTC m=+0.698724754,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.526089 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:07 crc kubenswrapper[4702]: E1007 12:24:07.580957 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.190:6443: connect: connection refused" interval="3.2s" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.684308 4702 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f272bee882c6210f9fe9f5af37df7965e8bd3e773e29027ca0ee53496e110c9d" exitCode=0 Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.684400 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f272bee882c6210f9fe9f5af37df7965e8bd3e773e29027ca0ee53496e110c9d"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.684479 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.685390 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.685420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.685430 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.692099 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.692184 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.693633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.693691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.693714 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.695799 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.695849 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.695865 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.695975 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.697694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.697730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.697740 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.698662 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.698708 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.699509 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.699541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.699553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.702931 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.702960 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.702975 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.702988 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9"} Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.910740 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.913006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.913044 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.913053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:07 crc kubenswrapper[4702]: I1007 12:24:07.913077 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:07 crc kubenswrapper[4702]: E1007 12:24:07.913819 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.190:6443: connect: connection refused" node="crc" Oct 07 12:24:07 crc kubenswrapper[4702]: W1007 12:24:07.947571 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:07 crc kubenswrapper[4702]: E1007 12:24:07.947686 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:08 crc kubenswrapper[4702]: W1007 12:24:08.047937 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:08 crc kubenswrapper[4702]: E1007 12:24:08.048052 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.153134 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:08 crc kubenswrapper[4702]: W1007 12:24:08.184066 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.190:6443: connect: connection refused Oct 07 12:24:08 crc kubenswrapper[4702]: E1007 12:24:08.184222 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.190:6443: connect: connection refused" logger="UnhandledError" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.710846 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a"} Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.711126 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.712788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.712858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.712884 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.715156 4702 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="011b7ff0a70a867cf2bbe100dbf1ebafe40b410576129e09dfa958014c0b3d86" exitCode=0 Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.715337 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.715374 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.715391 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"011b7ff0a70a867cf2bbe100dbf1ebafe40b410576129e09dfa958014c0b3d86"} Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.715499 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.716124 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.717847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.717959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.717992 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718162 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718226 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718356 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718778 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718878 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.718939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:08 crc kubenswrapper[4702]: I1007 12:24:08.923462 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724095 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724126 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724152 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724750 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b1f0cce1ebdc322c8367466f6396631e74f4991948ec81f4051512d398026c40"} Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724788 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef76a200b923258abceee8b6c914b1bbfff49fbe47f9ba9083e8e6201b908633"} Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724798 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"52e1ae7c219f810a11ad287256dfa460a2a85ee4873ef407410d1e9c7e7e6b7a"} Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724807 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b16bbf92ad960448a44a3547670cf7f14105e8fae8f1a359c53621c331cd304f"} Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.724971 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.725117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.725141 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.725152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.725994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.726047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.726064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.726005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.726162 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:09 crc kubenswrapper[4702]: I1007 12:24:09.726192 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.682164 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.732527 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef80396ff718cf571bd3f3ac0a4439ec2b73801e0763a79130bb5eda3c43a03f"} Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.732588 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.732638 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.734389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.734432 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.734446 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.735176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.735234 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:10 crc kubenswrapper[4702]: I1007 12:24:10.735250 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.114577 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.116571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.116653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.116673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.116722 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.222672 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.515445 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.735400 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.735401 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.736385 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.736414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.736422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.737041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.737073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:11 crc kubenswrapper[4702]: I1007 12:24:11.737091 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.422827 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.423214 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.425016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.425068 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.425083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.739240 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.739241 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.740760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.740808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.740822 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.740979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.741037 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.741058 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:12 crc kubenswrapper[4702]: I1007 12:24:12.865650 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.385792 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.386151 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.387969 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.388036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.388062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.396530 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.742627 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.743602 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.743687 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.746109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.746137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.746147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.746254 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.746316 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:13 crc kubenswrapper[4702]: I1007 12:24:13.746331 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:14 crc kubenswrapper[4702]: I1007 12:24:14.745621 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:14 crc kubenswrapper[4702]: I1007 12:24:14.747599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:14 crc kubenswrapper[4702]: I1007 12:24:14.747667 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:14 crc kubenswrapper[4702]: I1007 12:24:14.747686 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:14 crc kubenswrapper[4702]: E1007 12:24:14.801705 4702 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 12:24:15 crc kubenswrapper[4702]: I1007 12:24:15.423697 4702 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 12:24:15 crc kubenswrapper[4702]: I1007 12:24:15.423818 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 12:24:17 crc kubenswrapper[4702]: I1007 12:24:17.186206 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:17 crc kubenswrapper[4702]: I1007 12:24:17.186349 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:17 crc kubenswrapper[4702]: I1007 12:24:17.187412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:17 crc kubenswrapper[4702]: I1007 12:24:17.187442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:17 crc kubenswrapper[4702]: I1007 12:24:17.187451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:18 crc kubenswrapper[4702]: I1007 12:24:18.397891 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 07 12:24:18 crc kubenswrapper[4702]: I1007 12:24:18.398002 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 07 12:24:18 crc kubenswrapper[4702]: I1007 12:24:18.526550 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 07 12:24:19 crc kubenswrapper[4702]: W1007 12:24:19.087753 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 07 12:24:19 crc kubenswrapper[4702]: I1007 12:24:19.087859 4702 trace.go:236] Trace[1294252258]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 12:24:09.085) (total time: 10002ms): Oct 07 12:24:19 crc kubenswrapper[4702]: Trace[1294252258]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:24:19.087) Oct 07 12:24:19 crc kubenswrapper[4702]: Trace[1294252258]: [10.002062708s] [10.002062708s] END Oct 07 12:24:19 crc kubenswrapper[4702]: E1007 12:24:19.087880 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 07 12:24:19 crc kubenswrapper[4702]: I1007 12:24:19.360331 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 12:24:19 crc kubenswrapper[4702]: I1007 12:24:19.360412 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 12:24:19 crc kubenswrapper[4702]: I1007 12:24:19.364655 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 12:24:19 crc kubenswrapper[4702]: I1007 12:24:19.364770 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.691858 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.692637 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.695042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.695100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.695124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.697607 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.761883 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.762978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.763034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:20 crc kubenswrapper[4702]: I1007 12:24:20.763046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.261636 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.261842 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.263414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.263447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.263461 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.278753 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.765267 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.766996 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.767072 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:21 crc kubenswrapper[4702]: I1007 12:24:21.767098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.203998 4702 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.357373 4702 trace.go:236] Trace[1902812498]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 12:24:13.898) (total time: 10459ms): Oct 07 12:24:24 crc kubenswrapper[4702]: Trace[1902812498]: ---"Objects listed" error: 10459ms (12:24:24.357) Oct 07 12:24:24 crc kubenswrapper[4702]: Trace[1902812498]: [10.459070466s] [10.459070466s] END Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.357422 4702 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.360309 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.386196 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.392814 4702 trace.go:236] Trace[1149722051]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 12:24:14.244) (total time: 10148ms): Oct 07 12:24:24 crc kubenswrapper[4702]: Trace[1149722051]: ---"Objects listed" error: 10148ms (12:24:24.392) Oct 07 12:24:24 crc kubenswrapper[4702]: Trace[1149722051]: [10.148474343s] [10.148474343s] END Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.393238 4702 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.393108 4702 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.393354 4702 trace.go:236] Trace[193608710]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 12:24:13.042) (total time: 11350ms): Oct 07 12:24:24 crc kubenswrapper[4702]: Trace[193608710]: ---"Objects listed" error: 11350ms (12:24:24.392) Oct 07 12:24:24 crc kubenswrapper[4702]: Trace[193608710]: [11.350806302s] [11.350806302s] END Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.393534 4702 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.419340 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36712->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.419761 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36712->192.168.126.11:17697: read: connection reset by peer" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.419453 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36724->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.420976 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36724->192.168.126.11:17697: read: connection reset by peer" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.421580 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.421672 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.523395 4702 apiserver.go:52] "Watching apiserver" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.600279 4702 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.600566 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.600945 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.601031 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.601082 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.601350 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.602070 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.602092 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.602252 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.602351 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.602790 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.603559 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.603674 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.604019 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.604035 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.604060 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.604285 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.604732 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.605068 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.605847 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.639991 4702 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.655025 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.669395 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.690282 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696182 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696245 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696275 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696305 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696336 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696362 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696383 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696411 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696439 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696464 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696487 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696513 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696531 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696550 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696567 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696583 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696604 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696623 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696643 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696628 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696661 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696680 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696698 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696716 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696752 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696771 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696790 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696807 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696825 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696794 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696860 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696811 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696842 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696967 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.696996 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697020 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697038 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697042 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697127 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697154 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697179 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697201 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697220 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697238 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697263 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697279 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697296 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697311 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697330 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697355 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697376 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697392 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697414 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697430 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697448 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697465 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697481 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697472 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697497 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697513 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697521 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697528 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697551 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697595 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697595 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697625 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697656 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697676 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697698 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697716 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697735 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697740 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697754 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697774 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697797 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697794 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697819 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697840 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697876 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697930 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697959 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697970 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.697984 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698005 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698023 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698040 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698059 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698078 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698112 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698135 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698135 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698174 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698173 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698193 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698201 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698244 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698272 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698297 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698318 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698342 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698360 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698379 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698398 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698393 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698419 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698450 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698472 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698494 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698532 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698544 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698569 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698593 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698615 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698633 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698659 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698681 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698709 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698738 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698764 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698787 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698809 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698953 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.698957 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699020 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699041 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.699077 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:24:25.199046366 +0000 UTC m=+21.374782482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699106 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699134 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699164 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699166 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699189 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699223 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699236 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699251 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699282 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699312 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699322 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699339 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699376 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699411 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699421 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699441 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699472 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699479 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699502 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699596 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699657 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699692 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699719 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699738 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699748 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699794 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699797 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699819 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699853 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699879 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.699951 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700013 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700080 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700214 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700487 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700722 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700787 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700808 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.700988 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701035 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701094 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701122 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701147 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701171 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701206 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701232 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701255 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701279 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701310 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701334 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701359 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701350 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701385 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701443 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701482 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701521 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701549 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701579 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701605 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701632 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701661 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701689 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701717 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701746 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701770 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701795 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701820 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701847 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701861 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701873 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701919 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701945 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701950 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701969 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.701996 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702021 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702046 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702073 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702134 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702213 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702321 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702325 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702350 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702379 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702387 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702431 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702464 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702492 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702553 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702584 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702605 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702614 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702719 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702756 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702782 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702802 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702825 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702860 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702884 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702917 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702938 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702958 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.702987 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703008 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703026 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703046 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703068 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703090 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703115 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703136 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703155 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703176 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703197 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703217 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703238 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703260 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703316 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703352 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703377 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703397 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703420 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703443 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703475 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703501 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703846 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703950 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.703968 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705223 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705639 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705679 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705707 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705736 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705945 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705992 4702 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706095 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706119 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706136 4702 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706148 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706159 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706172 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706184 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706196 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706208 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706220 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706236 4702 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706250 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706264 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706276 4702 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706287 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706298 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706309 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706323 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706334 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706345 4702 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706358 4702 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706370 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706382 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706394 4702 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706409 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706419 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706430 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706440 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706451 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706462 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706473 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706484 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706495 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704160 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704286 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706506 4702 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.707202 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.707321 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.707356 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.707369 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.707410 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708047 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706847 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708184 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708204 4702 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708220 4702 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705922 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708235 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708256 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708269 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708305 4702 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708321 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708357 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708375 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708391 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704340 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704373 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704530 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704679 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704853 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.704951 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705143 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705142 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705180 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705446 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705475 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705535 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705654 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.705830 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706039 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706098 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.706110 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.708879 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709031 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709189 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709213 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709242 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709450 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709692 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709971 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.709997 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.710258 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.710406 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.710500 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.710864 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.711116 4702 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.713700 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.713839 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.714259 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.714268 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.714442 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.714737 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.714953 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:25.214829392 +0000 UTC m=+21.390565508 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.717924 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.719416 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.719684 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.719766 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:25.219746949 +0000 UTC m=+21.395483065 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.724296 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.725151 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.725370 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.725530 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.725992 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.726587 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.726832 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727088 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727264 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727279 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727303 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727385 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727776 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727786 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727829 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.727790 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.728035 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.728184 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.728374 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.728786 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.728932 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.729240 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.729988 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.733451 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.732507 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.741852 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.735257 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.733834 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.743006 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.733850 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.733889 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.734202 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.734729 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.743951 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.743973 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.744055 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:25.244033701 +0000 UTC m=+21.419769817 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.734836 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.744085 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.744137 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.744172 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:25.244161534 +0000 UTC m=+21.419897650 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.735125 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.735138 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.735500 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.735784 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.736157 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.736176 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.736547 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.736453 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.736835 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737132 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737365 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737350 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737421 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737499 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737730 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737760 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.737996 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.730022 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.744401 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.738142 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.738103 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.745089 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.745539 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.751956 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.752584 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.752775 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.753140 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.753586 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.754213 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.754417 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.755620 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.755137 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.760637 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.760658 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.760697 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761001 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761042 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761129 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761208 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761295 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761340 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761357 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761576 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761581 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761626 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761454 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761688 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.761826 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.762094 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.763284 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.763319 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.763549 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.763823 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.763845 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.763958 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.764605 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.764652 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.764800 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.765063 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.765086 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.765113 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.765346 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.765399 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770124 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770141 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770195 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770289 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770297 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770637 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770183 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770871 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.770969 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.771098 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.772229 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.773220 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.773433 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.775313 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.776114 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.781428 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.781705 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.784150 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.785909 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a" exitCode=255 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.786124 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a"} Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.789505 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.793547 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: E1007 12:24:24.795698 4702 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.802957 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.805439 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808803 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808847 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808943 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808958 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808968 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808980 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.808990 4702 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809070 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809207 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809541 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809564 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809576 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809587 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809597 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809607 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809642 4702 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809658 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809670 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809713 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809724 4702 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809734 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809742 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809750 4702 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809760 4702 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809770 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809779 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809788 4702 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809798 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809808 4702 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809818 4702 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809827 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809838 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809846 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809856 4702 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809865 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809874 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809883 4702 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809909 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809920 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809959 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809969 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.809978 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810006 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810015 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810025 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810038 4702 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810048 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810058 4702 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810068 4702 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810077 4702 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810087 4702 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810096 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810106 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810115 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810124 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810134 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810143 4702 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810152 4702 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810162 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810170 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810179 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810190 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810199 4702 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810189 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810209 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810288 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810307 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810321 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810333 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810344 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810355 4702 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810366 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810376 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810386 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810396 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810409 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810417 4702 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810426 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810435 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810444 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810453 4702 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810462 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810471 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810480 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810490 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810500 4702 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810509 4702 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810517 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810573 4702 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810583 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810593 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810602 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810611 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810623 4702 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810634 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810646 4702 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810654 4702 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810663 4702 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810672 4702 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810681 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810691 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810700 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810709 4702 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810719 4702 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810728 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810738 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810747 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810755 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810769 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810778 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810787 4702 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810796 4702 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810807 4702 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810815 4702 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810824 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810833 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810844 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810856 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810865 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810875 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810885 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810906 4702 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810920 4702 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810934 4702 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810945 4702 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810954 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.810990 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811003 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811013 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811023 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811033 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811043 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811053 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811064 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811075 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811085 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811095 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811105 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811115 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811125 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811135 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811146 4702 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811155 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811167 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811177 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811188 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811199 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811209 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811220 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.811230 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.816186 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.830142 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.838347 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.838927 4702 scope.go:117] "RemoveContainer" containerID="6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.839839 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.850883 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.865022 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.876657 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.887771 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.900764 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.912256 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.913280 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.918462 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.928956 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.930420 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: W1007 12:24:24.936680 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-79df8ec0a40ce811d7d097c87e2c593b43da88265c752f80b4e138edefe40ee0 WatchSource:0}: Error finding container 79df8ec0a40ce811d7d097c87e2c593b43da88265c752f80b4e138edefe40ee0: Status 404 returned error can't find the container with id 79df8ec0a40ce811d7d097c87e2c593b43da88265c752f80b4e138edefe40ee0 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.937387 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 12:24:24 crc kubenswrapper[4702]: W1007 12:24:24.940449 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6f8827e9476db0381c0edf80b4ce2f510dd330418cf71faddf968ffa071013ec WatchSource:0}: Error finding container 6f8827e9476db0381c0edf80b4ce2f510dd330418cf71faddf968ffa071013ec: Status 404 returned error can't find the container with id 6f8827e9476db0381c0edf80b4ce2f510dd330418cf71faddf968ffa071013ec Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.947104 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: W1007 12:24:24.950174 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-212551a0e86c452b556cbf745cd861a7df85f821ed0763d07018ad7c84a656c9 WatchSource:0}: Error finding container 212551a0e86c452b556cbf745cd861a7df85f821ed0763d07018ad7c84a656c9: Status 404 returned error can't find the container with id 212551a0e86c452b556cbf745cd861a7df85f821ed0763d07018ad7c84a656c9 Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.961063 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.973837 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.987136 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:24 crc kubenswrapper[4702]: I1007 12:24:24.998766 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.011536 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.023418 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.037459 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.051069 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.062720 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.072749 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.214651 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.214943 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:24:26.21488945 +0000 UTC m=+22.390625576 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.315866 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.315947 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.315978 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.316009 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.316150 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.316170 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.316184 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.316244 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:26.316227275 +0000 UTC m=+22.491963391 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.316705 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.316744 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:26.316736299 +0000 UTC m=+22.492472415 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.317014 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.317042 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.317156 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:26.31712395 +0000 UTC m=+22.492860066 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.317164 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.317176 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:25 crc kubenswrapper[4702]: E1007 12:24:25.317208 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:26.317198172 +0000 UTC m=+22.492934288 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.790280 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"212551a0e86c452b556cbf745cd861a7df85f821ed0763d07018ad7c84a656c9"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.792590 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.792736 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.792774 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6f8827e9476db0381c0edf80b4ce2f510dd330418cf71faddf968ffa071013ec"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.795673 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.795712 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"79df8ec0a40ce811d7d097c87e2c593b43da88265c752f80b4e138edefe40ee0"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.798134 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.799860 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b"} Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.799959 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.807105 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.821856 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.847800 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.863156 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.875919 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.890080 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.907044 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.922768 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.938423 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.953892 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.969137 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:25 crc kubenswrapper[4702]: I1007 12:24:25.986773 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.001985 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:26Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.019066 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:26Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.033870 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:26Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.050022 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:26Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.224080 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.224423 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:24:28.224375677 +0000 UTC m=+24.400111793 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.324999 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.325074 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.325121 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.325157 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325218 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325344 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325380 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325395 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325442 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325458 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325403 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325358 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:28.325325851 +0000 UTC m=+24.501062107 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325395 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325596 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:28.325567278 +0000 UTC m=+24.501303394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325617 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:28.325607699 +0000 UTC m=+24.501344035 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.325731 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:28.325691021 +0000 UTC m=+24.501427277 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.640045 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.640206 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.640340 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.640369 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.640551 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:26 crc kubenswrapper[4702]: E1007 12:24:26.640710 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.644615 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.645496 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.646786 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.647405 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.648403 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.648951 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.649537 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.731128 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.731863 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.733206 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.733744 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.734983 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.735496 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.760721 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.761397 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.762511 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.763299 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.764235 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.764953 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.765675 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.766766 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.767360 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.767820 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.769745 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.771296 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.772585 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.773970 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.774938 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.775889 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.776656 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.777564 4702 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.777723 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.782301 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.783942 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.784862 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.787706 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.789237 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.790420 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.792623 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.795123 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.796462 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.798598 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.799823 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.801065 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.802012 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.804377 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.806939 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.808636 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.809873 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.811994 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.813027 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.815373 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.816706 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 07 12:24:26 crc kubenswrapper[4702]: I1007 12:24:26.818012 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.807625 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3"} Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.831705 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.846673 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.859663 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.873112 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.890791 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.908309 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.925172 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:27 crc kubenswrapper[4702]: I1007 12:24:27.941754 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:27Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.244076 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.244282 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:24:32.244249316 +0000 UTC m=+28.419985442 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.344756 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.344875 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.345006 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.345064 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345142 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345257 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:32.345235861 +0000 UTC m=+28.520971977 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345299 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345138 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345409 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345420 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:32.345382515 +0000 UTC m=+28.521118851 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345437 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345573 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:32.345529449 +0000 UTC m=+28.521265725 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345693 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345710 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345724 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.345791 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:32.345781106 +0000 UTC m=+28.521517212 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.640462 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.640529 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:28 crc kubenswrapper[4702]: I1007 12:24:28.640482 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.640697 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.640807 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:28 crc kubenswrapper[4702]: E1007 12:24:28.640980 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.906229 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-djf9f"] Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.906593 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-t9vtm"] Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.906836 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.907558 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.914135 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-h9s7h"] Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.914601 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9s7h" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.922479 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.922648 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.922654 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.923223 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.923378 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.924258 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.924340 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.926570 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.927464 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.936774 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.961192 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.978641 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:29 crc kubenswrapper[4702]: I1007 12:24:29.993252 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.004706 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.020359 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.033888 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.050941 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.057923 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqgzf\" (UniqueName: \"kubernetes.io/projected/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-kube-api-access-mqgzf\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.057962 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-cni-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.057980 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-daemon-config\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058000 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f302f84b-82f7-4c73-a179-e7c4e20d05c5-hosts-file\") pod \"node-resolver-djf9f\" (UID: \"f302f84b-82f7-4c73-a179-e7c4e20d05c5\") " pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058031 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/943865b8-ec83-4c7e-9e71-b02eaea71c1d-cni-binary-copy\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058184 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-cni-multus\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058325 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058381 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-cnibin\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058445 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-k8s-cni-cncf-io\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058491 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-kubelet\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058530 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-hostroot\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058593 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cnibin\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058641 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7h6p\" (UniqueName: \"kubernetes.io/projected/f302f84b-82f7-4c73-a179-e7c4e20d05c5-kube-api-access-d7h6p\") pod \"node-resolver-djf9f\" (UID: \"f302f84b-82f7-4c73-a179-e7c4e20d05c5\") " pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058662 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-system-cni-dir\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058684 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5xqg\" (UniqueName: \"kubernetes.io/projected/943865b8-ec83-4c7e-9e71-b02eaea71c1d-kube-api-access-m5xqg\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058708 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-os-release\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058738 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-conf-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058792 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-socket-dir-parent\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.058815 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-etc-kubernetes\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059029 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cni-binary-copy\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059102 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059129 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-system-cni-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059155 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-os-release\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059179 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-netns\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059237 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-cni-bin\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.059267 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-multus-certs\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.064153 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.077289 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.095066 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.112718 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.128035 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.143369 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.157933 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160441 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-conf-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160494 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-socket-dir-parent\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160520 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-etc-kubernetes\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160568 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cni-binary-copy\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160598 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160633 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-system-cni-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160660 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-os-release\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160656 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-conf-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160709 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-etc-kubernetes\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160741 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-netns\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160688 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-netns\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160675 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-socket-dir-parent\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160736 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-system-cni-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160821 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-cni-bin\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160848 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-multus-certs\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160874 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqgzf\" (UniqueName: \"kubernetes.io/projected/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-kube-api-access-mqgzf\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160916 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-daemon-config\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160934 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f302f84b-82f7-4c73-a179-e7c4e20d05c5-hosts-file\") pod \"node-resolver-djf9f\" (UID: \"f302f84b-82f7-4c73-a179-e7c4e20d05c5\") " pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160929 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-cni-bin\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160969 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-cni-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160935 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-multus-certs\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.160987 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/943865b8-ec83-4c7e-9e71-b02eaea71c1d-cni-binary-copy\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161098 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-os-release\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161122 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-cni-multus\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161161 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-cni-multus\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161214 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f302f84b-82f7-4c73-a179-e7c4e20d05c5-hosts-file\") pod \"node-resolver-djf9f\" (UID: \"f302f84b-82f7-4c73-a179-e7c4e20d05c5\") " pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161268 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161324 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-cnibin\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161364 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-k8s-cni-cncf-io\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161392 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-kubelet\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161384 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-cni-dir\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161419 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-hostroot\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161448 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-run-k8s-cni-cncf-io\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161456 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-hostroot\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161468 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-cnibin\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161496 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cnibin\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161502 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/943865b8-ec83-4c7e-9e71-b02eaea71c1d-host-var-lib-kubelet\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161530 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-system-cni-dir\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161551 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5xqg\" (UniqueName: \"kubernetes.io/projected/943865b8-ec83-4c7e-9e71-b02eaea71c1d-kube-api-access-m5xqg\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161556 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cnibin\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161573 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7h6p\" (UniqueName: \"kubernetes.io/projected/f302f84b-82f7-4c73-a179-e7c4e20d05c5-kube-api-access-d7h6p\") pod \"node-resolver-djf9f\" (UID: \"f302f84b-82f7-4c73-a179-e7c4e20d05c5\") " pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161591 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-os-release\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161594 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-system-cni-dir\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161680 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-os-release\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161688 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cni-binary-copy\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161830 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/943865b8-ec83-4c7e-9e71-b02eaea71c1d-cni-binary-copy\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.161910 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/943865b8-ec83-4c7e-9e71-b02eaea71c1d-multus-daemon-config\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.162024 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.162397 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.171814 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.179640 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqgzf\" (UniqueName: \"kubernetes.io/projected/2bf8728e-da3d-4d2a-9b9b-6094b460f50b-kube-api-access-mqgzf\") pod \"multus-additional-cni-plugins-t9vtm\" (UID: \"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\") " pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.179938 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5xqg\" (UniqueName: \"kubernetes.io/projected/943865b8-ec83-4c7e-9e71-b02eaea71c1d-kube-api-access-m5xqg\") pod \"multus-h9s7h\" (UID: \"943865b8-ec83-4c7e-9e71-b02eaea71c1d\") " pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.182215 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7h6p\" (UniqueName: \"kubernetes.io/projected/f302f84b-82f7-4c73-a179-e7c4e20d05c5-kube-api-access-d7h6p\") pod \"node-resolver-djf9f\" (UID: \"f302f84b-82f7-4c73-a179-e7c4e20d05c5\") " pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.186617 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.205376 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.221247 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.222209 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.232974 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-djf9f" Oct 07 12:24:30 crc kubenswrapper[4702]: W1007 12:24:30.237819 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bf8728e_da3d_4d2a_9b9b_6094b460f50b.slice/crio-2ab426a48a18417395a11285f51256d2e276e45ee47cc57a1845c841c866ba71 WatchSource:0}: Error finding container 2ab426a48a18417395a11285f51256d2e276e45ee47cc57a1845c841c866ba71: Status 404 returned error can't find the container with id 2ab426a48a18417395a11285f51256d2e276e45ee47cc57a1845c841c866ba71 Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.240563 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9s7h" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.244632 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.262298 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.318735 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-w4btn"] Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.319289 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pcg5c"] Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.319561 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.320793 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.321927 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.321991 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.323379 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.323449 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.323549 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.323639 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.323845 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.324143 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.324304 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.324324 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.324335 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.324954 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.339921 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.358238 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.371932 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.389521 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.405522 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.436700 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.457880 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464747 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-slash\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464797 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-var-lib-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464824 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e1ca41d2-75d4-4871-abe5-46a3e172fadd-rootfs\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464844 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1ca41d2-75d4-4871-abe5-46a3e172fadd-mcd-auth-proxy-config\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464867 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-node-log\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464885 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1ca41d2-75d4-4871-abe5-46a3e172fadd-proxy-tls\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.464932 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465073 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-log-socket\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465131 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-config\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465174 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-kubelet\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465201 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-systemd\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465223 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-ovn-kubernetes\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465256 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8ckk\" (UniqueName: \"kubernetes.io/projected/e1ca41d2-75d4-4871-abe5-46a3e172fadd-kube-api-access-g8ckk\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465285 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d2c9\" (UniqueName: \"kubernetes.io/projected/68b62ec2-6971-4e45-bff9-5b65b657b6b6-kube-api-access-6d2c9\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465326 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-netns\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465356 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-env-overrides\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465380 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-netd\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465407 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465430 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovn-node-metrics-cert\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465453 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-systemd-units\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465478 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-etc-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465502 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-bin\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465533 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-ovn\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.465578 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-script-lib\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.477589 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.494865 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.515735 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.528258 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.543769 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.559702 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566456 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-log-socket\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566501 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566553 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-config\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566584 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-ovn-kubernetes\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566589 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-log-socket\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566614 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-kubelet\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566649 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-systemd\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566681 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8ckk\" (UniqueName: \"kubernetes.io/projected/e1ca41d2-75d4-4871-abe5-46a3e172fadd-kube-api-access-g8ckk\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566657 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566706 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d2c9\" (UniqueName: \"kubernetes.io/projected/68b62ec2-6971-4e45-bff9-5b65b657b6b6-kube-api-access-6d2c9\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566784 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-netns\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566861 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-env-overrides\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566887 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-netd\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566926 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566944 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovn-node-metrics-cert\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566962 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-bin\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.566983 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-systemd-units\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567001 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-etc-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567019 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-ovn\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567030 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-ovn-kubernetes\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567050 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-script-lib\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567079 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-kubelet\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567088 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-slash\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567114 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-systemd\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567148 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-netns\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567150 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e1ca41d2-75d4-4871-abe5-46a3e172fadd-rootfs\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567116 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e1ca41d2-75d4-4871-abe5-46a3e172fadd-rootfs\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567328 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1ca41d2-75d4-4871-abe5-46a3e172fadd-mcd-auth-proxy-config\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567354 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-var-lib-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567373 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-node-log\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567393 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1ca41d2-75d4-4871-abe5-46a3e172fadd-proxy-tls\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567501 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-config\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567549 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-var-lib-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567691 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-systemd-units\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567731 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567891 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-node-log\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567765 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-netd\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567771 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-etc-openvswitch\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567787 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-ovn\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.567747 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-env-overrides\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.568048 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-slash\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.568270 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-bin\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.568274 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1ca41d2-75d4-4871-abe5-46a3e172fadd-mcd-auth-proxy-config\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.568378 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-script-lib\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.572680 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovn-node-metrics-cert\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.573188 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1ca41d2-75d4-4871-abe5-46a3e172fadd-proxy-tls\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.574915 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.585949 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d2c9\" (UniqueName: \"kubernetes.io/projected/68b62ec2-6971-4e45-bff9-5b65b657b6b6-kube-api-access-6d2c9\") pod \"ovnkube-node-pcg5c\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.586493 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8ckk\" (UniqueName: \"kubernetes.io/projected/e1ca41d2-75d4-4871-abe5-46a3e172fadd-kube-api-access-g8ckk\") pod \"machine-config-daemon-w4btn\" (UID: \"e1ca41d2-75d4-4871-abe5-46a3e172fadd\") " pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.588889 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.608161 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.621084 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.634696 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.640748 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.640797 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.640870 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.640915 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.641056 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.641223 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.645094 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.655122 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: W1007 12:24:30.659178 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1ca41d2_75d4_4871_abe5_46a3e172fadd.slice/crio-2d198e60e9d3ab389d731b055b12930af0ca0a4081da5c576ebd0115dece5830 WatchSource:0}: Error finding container 2d198e60e9d3ab389d731b055b12930af0ca0a4081da5c576ebd0115dece5830: Status 404 returned error can't find the container with id 2d198e60e9d3ab389d731b055b12930af0ca0a4081da5c576ebd0115dece5830 Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.673673 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:30 crc kubenswrapper[4702]: W1007 12:24:30.713980 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68b62ec2_6971_4e45_bff9_5b65b657b6b6.slice/crio-3aa3931da2c247feeb5f4a2682d04beaac9294b0792af0a544809f6b69162c8c WatchSource:0}: Error finding container 3aa3931da2c247feeb5f4a2682d04beaac9294b0792af0a544809f6b69162c8c: Status 404 returned error can't find the container with id 3aa3931da2c247feeb5f4a2682d04beaac9294b0792af0a544809f6b69162c8c Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.739370 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.764737 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.781025 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.786581 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.788750 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.788802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.788814 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.789031 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.794437 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.796344 4702 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.796434 4702 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.797532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.797581 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.797591 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.797611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.797622 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.809681 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.815853 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.816618 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"3aa3931da2c247feeb5f4a2682d04beaac9294b0792af0a544809f6b69162c8c"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.818193 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-djf9f" event={"ID":"f302f84b-82f7-4c73-a179-e7c4e20d05c5","Type":"ContainerStarted","Data":"a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.818268 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-djf9f" event={"ID":"f302f84b-82f7-4c73-a179-e7c4e20d05c5","Type":"ContainerStarted","Data":"791f7e174f7607dcae4a3a4a147d1f167a14d9b648b8570b7e1c6a2dc7b59806"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.819548 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"2d198e60e9d3ab389d731b055b12930af0ca0a4081da5c576ebd0115dece5830"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.820803 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.820832 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.820842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.820859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.820873 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.826317 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerStarted","Data":"8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.826388 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerStarted","Data":"4d894551958b8978e350c42e8ebe56fd661e53106ae778f7ba4d7534465620de"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.828093 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.829205 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerStarted","Data":"f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.829260 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerStarted","Data":"2ab426a48a18417395a11285f51256d2e276e45ee47cc57a1845c841c866ba71"} Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.837135 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.843026 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.843061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.843071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.843092 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.843109 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.843046 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.854363 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.855757 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.858640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.858696 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.858773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.858806 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.858825 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.870723 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.874737 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.878335 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.878368 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.878378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.878395 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.878407 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.882437 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.891262 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: E1007 12:24:30.891464 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.893106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.893130 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.893137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.893153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.893163 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.896840 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.915516 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.930312 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.944881 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.958554 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.976149 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.992187 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.996427 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.996492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.996510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.996535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:30 crc kubenswrapper[4702]: I1007 12:24:30.996553 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:30Z","lastTransitionTime":"2025-10-07T12:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.006762 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.021591 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.039565 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.054222 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.069584 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.082414 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.099059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.099115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.099126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.099148 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.099163 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.102992 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.118547 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.133985 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.147097 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.177515 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.202003 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.202070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.202080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.202098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.202110 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.217910 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.261701 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.298591 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.304843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.304890 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.304924 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.304944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.304958 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.339690 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.408408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.408464 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.408476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.408497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.408510 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.511135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.511186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.511201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.511219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.511232 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.614644 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.614704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.614713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.614734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.614747 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.717668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.717744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.717760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.717792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.717811 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.820704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.820757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.820768 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.820789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.820803 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.838622 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.838687 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.841649 4702 generic.go:334] "Generic (PLEG): container finished" podID="2bf8728e-da3d-4d2a-9b9b-6094b460f50b" containerID="f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90" exitCode=0 Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.841707 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerDied","Data":"f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.844488 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" exitCode=0 Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.844543 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.859183 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.879829 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.901829 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.913721 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.924478 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.925135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.925150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.925176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.925185 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:31Z","lastTransitionTime":"2025-10-07T12:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.930999 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.954984 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.971508 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:31 crc kubenswrapper[4702]: I1007 12:24:31.988321 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.004578 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.023515 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.028532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.028589 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.028603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.028623 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.028637 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.038010 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.054970 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.074373 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.091539 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.107144 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.125859 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.130844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.130868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.130877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.130896 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.130942 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.141830 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.161129 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.178934 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.191558 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.207940 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.221484 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.233544 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.233593 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.233606 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.233625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.233638 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.259142 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.284689 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.284968 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:24:40.284942069 +0000 UTC m=+36.460678185 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.296307 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.336849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.336925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.336940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.336967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.336980 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.341934 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.385782 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.385837 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.385868 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.385885 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386006 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386025 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386037 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386044 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386006 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386091 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:40.386072948 +0000 UTC m=+36.561809064 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386093 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386105 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:40.386100408 +0000 UTC m=+36.561836524 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386105 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386121 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386136 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:40.386127319 +0000 UTC m=+36.561863435 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.386272 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:40.386241362 +0000 UTC m=+36.561977488 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.391353 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.440110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.440174 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.440200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.440225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.440242 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.544135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.544626 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.544637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.544656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.544678 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.640201 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.640247 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.640278 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.640367 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.640505 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:32 crc kubenswrapper[4702]: E1007 12:24:32.640597 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.646476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.646511 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.646521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.646537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.646557 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.749255 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.749310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.749323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.749344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.749357 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.850999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.851043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.851055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.851014 4702 generic.go:334] "Generic (PLEG): container finished" podID="2bf8728e-da3d-4d2a-9b9b-6094b460f50b" containerID="695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d" exitCode=0 Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.851071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.851083 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.851042 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerDied","Data":"695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.853679 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.853705 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.853719 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.853733 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.870525 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.889663 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.903661 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.920804 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.937914 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.951473 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.953310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.953354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.953365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.953386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.953399 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:32Z","lastTransitionTime":"2025-10-07T12:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.968780 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.984806 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2jrz4"] Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.985824 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.987955 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.988185 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.988472 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.988756 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 07 12:24:32 crc kubenswrapper[4702]: I1007 12:24:32.991329 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.004452 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.022005 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.034656 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.047307 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.057214 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.057272 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.057286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.057309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.057323 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.061751 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.073684 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.083370 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.095988 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1d66a83d-539e-4a9b-8299-bf79fcd54629-serviceca\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.096030 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4htvd\" (UniqueName: \"kubernetes.io/projected/1d66a83d-539e-4a9b-8299-bf79fcd54629-kube-api-access-4htvd\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.096216 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d66a83d-539e-4a9b-8299-bf79fcd54629-host\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.101812 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.141731 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.159616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.159658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.159670 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.159687 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.159700 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.179508 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.196865 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1d66a83d-539e-4a9b-8299-bf79fcd54629-serviceca\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.196942 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4htvd\" (UniqueName: \"kubernetes.io/projected/1d66a83d-539e-4a9b-8299-bf79fcd54629-kube-api-access-4htvd\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.196969 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d66a83d-539e-4a9b-8299-bf79fcd54629-host\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.197054 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d66a83d-539e-4a9b-8299-bf79fcd54629-host\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.198070 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1d66a83d-539e-4a9b-8299-bf79fcd54629-serviceca\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.219189 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.252372 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4htvd\" (UniqueName: \"kubernetes.io/projected/1d66a83d-539e-4a9b-8299-bf79fcd54629-kube-api-access-4htvd\") pod \"node-ca-2jrz4\" (UID: \"1d66a83d-539e-4a9b-8299-bf79fcd54629\") " pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.263732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.263781 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.263792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.263812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.263824 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.287988 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.299827 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jrz4" Oct 07 12:24:33 crc kubenswrapper[4702]: W1007 12:24:33.318288 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d66a83d_539e_4a9b_8299_bf79fcd54629.slice/crio-fde49350df93bf52cee88bdd0706255020e7fa742c898ba057f52c762b7d3261 WatchSource:0}: Error finding container fde49350df93bf52cee88bdd0706255020e7fa742c898ba057f52c762b7d3261: Status 404 returned error can't find the container with id fde49350df93bf52cee88bdd0706255020e7fa742c898ba057f52c762b7d3261 Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.323736 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.359360 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.367287 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.367344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.367355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.367374 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.367388 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.403125 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.445594 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.472154 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.472197 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.472210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.472230 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.472270 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.481218 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.522161 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.561222 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.578891 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.579006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.579027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.579061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.579087 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.682041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.682887 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.683006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.683110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.683196 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.786223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.786266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.786276 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.786293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.786303 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.865422 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.865482 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.867992 4702 generic.go:334] "Generic (PLEG): container finished" podID="2bf8728e-da3d-4d2a-9b9b-6094b460f50b" containerID="d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda" exitCode=0 Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.868046 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerDied","Data":"d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.872049 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jrz4" event={"ID":"1d66a83d-539e-4a9b-8299-bf79fcd54629","Type":"ContainerStarted","Data":"324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.872115 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jrz4" event={"ID":"1d66a83d-539e-4a9b-8299-bf79fcd54629","Type":"ContainerStarted","Data":"fde49350df93bf52cee88bdd0706255020e7fa742c898ba057f52c762b7d3261"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.888625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.888668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.888678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.888699 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.888713 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.892039 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.910848 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.924341 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.935862 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.959221 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.973506 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.987679 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.992959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.993010 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.993020 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.993042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:33 crc kubenswrapper[4702]: I1007 12:24:33.993054 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:33Z","lastTransitionTime":"2025-10-07T12:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.001604 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:33Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.014440 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.031247 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.049061 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.064025 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.080119 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.095577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.095831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.095943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.096061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.096149 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.118196 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.159518 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.198494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.198528 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.198536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.198554 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.198564 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.198760 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.241398 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.279278 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.301299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.301380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.301395 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.301417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.301449 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.321037 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.359271 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.399102 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.404580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.404629 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.404640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.404661 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.404671 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.439103 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.482041 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.507415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.507482 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.507502 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.507533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.507552 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.524057 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.561421 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.600052 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.610657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.610729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.610753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.610785 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.610811 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.640927 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.640957 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.640957 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:34 crc kubenswrapper[4702]: E1007 12:24:34.641112 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:34 crc kubenswrapper[4702]: E1007 12:24:34.641237 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:34 crc kubenswrapper[4702]: E1007 12:24:34.641437 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.642653 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.678310 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.713590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.713618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.713629 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.713644 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.713655 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.721400 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.759310 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.799109 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.816584 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.816639 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.816655 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.816676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.816688 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.840055 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.879357 4702 generic.go:334] "Generic (PLEG): container finished" podID="2bf8728e-da3d-4d2a-9b9b-6094b460f50b" containerID="8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81" exitCode=0 Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.879411 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerDied","Data":"8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.883537 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.919632 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.919861 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.919924 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.919943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.919965 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.919979 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:34Z","lastTransitionTime":"2025-10-07T12:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.960743 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:34 crc kubenswrapper[4702]: I1007 12:24:34.998631 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.022998 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.023069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.023087 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.023116 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.023135 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.041208 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.086576 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.120422 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.126707 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.126739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.126761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.126783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.126796 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.159301 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.198857 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.229921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.229961 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.229971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.229986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.229997 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.236642 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.285682 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.322994 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.333290 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.333338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.333353 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.333384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.333400 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.360316 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.403573 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.436582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.436687 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.436713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.436748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.436771 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.447447 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.478964 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.521411 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.539739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.539789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.539804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.539823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.539836 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.558819 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.602347 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.642253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.642340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.642359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.642389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.642409 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.644120 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.680091 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.730421 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.744829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.744879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.744909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.744932 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.744950 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.762347 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.801494 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.847466 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.847531 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.847549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.847572 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.847584 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.888739 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerStarted","Data":"d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.912021 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.941178 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.949753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.949794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.949803 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.949818 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.949830 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:35Z","lastTransitionTime":"2025-10-07T12:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.962863 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:35 crc kubenswrapper[4702]: I1007 12:24:35.979785 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:35Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.003770 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.046368 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.056591 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.056660 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.056675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.056699 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.056715 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.084990 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.116983 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.157866 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.159179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.159216 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.159228 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.159252 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.159268 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.202366 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.240487 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.262706 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.262749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.262760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.262779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.262792 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.279969 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.320337 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.357496 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.366152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.366384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.366398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.366419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.366432 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.469289 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.469344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.469353 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.469373 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.469387 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.571642 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.571715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.571751 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.571777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.571791 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.640452 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:36 crc kubenswrapper[4702]: E1007 12:24:36.640629 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.640461 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:36 crc kubenswrapper[4702]: E1007 12:24:36.641120 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.641114 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:36 crc kubenswrapper[4702]: E1007 12:24:36.641214 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.674296 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.674355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.674364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.674394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.674406 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.777309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.777359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.777372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.777393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.777409 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.880374 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.880420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.880435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.880457 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.880474 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.895733 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.898833 4702 generic.go:334] "Generic (PLEG): container finished" podID="2bf8728e-da3d-4d2a-9b9b-6094b460f50b" containerID="d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd" exitCode=0 Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.898868 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerDied","Data":"d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.913123 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.928095 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.959339 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.974748 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.983823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.983881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.983930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.983951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.983963 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:36Z","lastTransitionTime":"2025-10-07T12:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:36 crc kubenswrapper[4702]: I1007 12:24:36.992600 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:36Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.006127 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.018501 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.033616 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.050019 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.065407 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.083389 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.086047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.086088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.086101 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.086120 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.086133 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.103976 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.122236 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.142450 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.188915 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.188952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.188961 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.188977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.188986 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.292616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.292677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.292694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.292717 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.292735 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.396558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.396658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.396688 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.396724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.396744 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.500286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.500343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.500359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.500384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.500398 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.604487 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.604549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.604563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.604586 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.604604 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.706992 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.707050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.707058 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.707105 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.707116 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.810444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.810495 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.810506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.810526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.810544 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.908020 4702 generic.go:334] "Generic (PLEG): container finished" podID="2bf8728e-da3d-4d2a-9b9b-6094b460f50b" containerID="9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af" exitCode=0 Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.908149 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerDied","Data":"9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.912248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.912313 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.912331 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.912358 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.912378 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:37Z","lastTransitionTime":"2025-10-07T12:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.927259 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.938562 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.952470 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.969294 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:37 crc kubenswrapper[4702]: I1007 12:24:37.983533 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:37Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.005014 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.014845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.014918 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.014937 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.014959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.014975 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.024422 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.040061 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.055519 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.080232 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.094517 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.116847 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.118078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.118157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.118176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.118211 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.118236 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.142727 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.165443 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.220939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.221001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.221010 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.221028 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.221039 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.323849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.323924 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.323938 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.323961 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.323974 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.403077 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.421560 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.427199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.427280 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.427308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.427343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.427371 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.443114 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.457011 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.475006 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.500389 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.520144 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.531125 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.531169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.531182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.531203 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.531220 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.538123 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.555424 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.575254 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.590286 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.604611 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.619397 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.633485 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.635305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.635340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.635352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.635375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.635392 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.640113 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.640237 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:38 crc kubenswrapper[4702]: E1007 12:24:38.640245 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.640407 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:38 crc kubenswrapper[4702]: E1007 12:24:38.640576 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:38 crc kubenswrapper[4702]: E1007 12:24:38.640746 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.667828 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.740697 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.740755 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.740777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.740803 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.740822 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.843191 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.843248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.843261 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.843283 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.843301 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.920682 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.921190 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.921321 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.921366 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.927672 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" event={"ID":"2bf8728e-da3d-4d2a-9b9b-6094b460f50b","Type":"ContainerStarted","Data":"dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.939812 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.946102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.946166 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.946184 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.946211 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.946230 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:38Z","lastTransitionTime":"2025-10-07T12:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.952867 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.954543 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.956826 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.974243 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:38 crc kubenswrapper[4702]: I1007 12:24:38.991954 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:38Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.004923 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.021533 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.044756 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.049314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.049371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.049381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.049397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.049407 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.063576 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.080262 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.096961 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.114119 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.129411 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.148392 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.152856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.152939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.152955 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.152975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.152988 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.164455 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.175305 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.189320 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.214158 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.228593 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.245284 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.255708 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.255798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.255808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.255829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.255839 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.261961 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.277017 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.290163 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.308450 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.322437 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.335357 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.348722 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.357939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.358159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.358239 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.358345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.358440 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.363926 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.377659 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:39Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.461035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.461319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.461397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.461488 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.461557 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.564407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.564734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.564803 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.564882 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.565003 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.668598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.668949 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.669046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.669177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.669274 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.773144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.773241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.773274 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.773340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.773365 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.876572 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.876675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.876701 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.876734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.876764 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.980629 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.980676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.980694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.980714 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:39 crc kubenswrapper[4702]: I1007 12:24:39.980726 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:39Z","lastTransitionTime":"2025-10-07T12:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.083946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.083986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.083997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.084014 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.084025 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.186894 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.187026 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.187049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.187081 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.187112 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.289588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.289655 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.289673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.289700 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.289719 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.375002 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.375613 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:24:56.375566962 +0000 UTC m=+52.551303098 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.392340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.392393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.392407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.392425 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.392437 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.476400 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.476445 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.476477 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.476505 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476612 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476644 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476762 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476830 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476845 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476782 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:56.476746392 +0000 UTC m=+52.652482548 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476768 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476970 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476993 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.476942 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:56.476916547 +0000 UTC m=+52.652652663 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.477045 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:56.47703517 +0000 UTC m=+52.652771286 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.477083 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:56.47705154 +0000 UTC m=+52.652787696 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.495381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.495443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.495460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.495487 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.495510 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.598550 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.598618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.598645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.598680 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.598705 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.640630 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.640726 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.640870 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.641034 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.643774 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:40 crc kubenswrapper[4702]: E1007 12:24:40.644023 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.704077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.704613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.704636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.704660 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.704677 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.808028 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.808086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.808107 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.808142 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.808163 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.911496 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.911558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.911576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.911611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:40 crc kubenswrapper[4702]: I1007 12:24:40.911629 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:40Z","lastTransitionTime":"2025-10-07T12:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.014281 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.014352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.014367 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.014392 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.014412 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.117306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.117379 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.117397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.117426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.117449 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.219974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.220039 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.220050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.220070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.220084 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.282266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.282327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.282348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.282375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.282394 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: E1007 12:24:41.296717 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:41Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.301191 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.301225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.301234 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.301250 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.301262 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: E1007 12:24:41.312169 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:41Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.316826 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.316930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.316950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.316979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.317002 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: E1007 12:24:41.331226 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:41Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.335836 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.335960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.335999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.336034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.336060 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: E1007 12:24:41.348252 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:41Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.353274 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.353314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.353326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.353364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.353379 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: E1007 12:24:41.367838 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:41Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:41 crc kubenswrapper[4702]: E1007 12:24:41.368018 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.370150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.370207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.370222 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.370242 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.370256 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.473065 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.473132 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.473144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.473165 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.473177 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.577087 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.577158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.577172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.577195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.577210 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.680496 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.680561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.680594 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.680617 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.680639 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.782999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.783049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.783061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.783078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.783093 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.886174 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.886246 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.886265 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.886298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.886321 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.992558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.992619 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.993327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.993388 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:41 crc kubenswrapper[4702]: I1007 12:24:41.993413 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:41Z","lastTransitionTime":"2025-10-07T12:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.097454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.097492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.097503 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.097522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.097532 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.200274 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.200323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.200332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.200348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.200358 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.303121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.303522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.303631 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.303770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.303845 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.407391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.407815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.408040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.408246 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.408416 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.512409 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.512487 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.512508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.512539 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.512565 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.615532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.615615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.615640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.615672 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.615700 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.641101 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.641162 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.641124 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:42 crc kubenswrapper[4702]: E1007 12:24:42.641355 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:42 crc kubenswrapper[4702]: E1007 12:24:42.641523 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:42 crc kubenswrapper[4702]: E1007 12:24:42.641667 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.719403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.719499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.719520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.719544 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.719559 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.822754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.822817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.822833 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.822860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.822882 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.851065 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9"] Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.851755 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.854395 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.855445 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.876554 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.895037 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.907457 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xxh8\" (UniqueName: \"kubernetes.io/projected/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-kube-api-access-8xxh8\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.907659 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.907820 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.907956 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.916390 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.925949 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.925992 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.926004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.926030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.926045 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:42Z","lastTransitionTime":"2025-10-07T12:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.937663 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.951408 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/0.log" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.955351 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698" exitCode=1 Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.955436 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698"} Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.956369 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.957214 4702 scope.go:117] "RemoveContainer" containerID="2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.973323 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:42 crc kubenswrapper[4702]: I1007 12:24:42.986703 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:42Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.009275 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.009355 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.009419 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xxh8\" (UniqueName: \"kubernetes.io/projected/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-kube-api-access-8xxh8\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.009503 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.009672 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.010635 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.010837 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.018335 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.029445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.029542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.029568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.029617 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.029643 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.045499 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xxh8\" (UniqueName: \"kubernetes.io/projected/ad6b96be-e641-4d3e-a8d8-a750d029fdc5-kube-api-access-8xxh8\") pod \"ovnkube-control-plane-749d76644c-f4cq9\" (UID: \"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.046998 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.065614 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.079166 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.131183 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.133113 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.133140 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.133149 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.133170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.133181 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.164725 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.178125 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.183257 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.205339 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.217345 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.230027 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.237946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.237987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.237998 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.238017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.238029 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.245951 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.260814 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.271222 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.284806 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.305064 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091824 6070 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 12:24:42.091833 6070 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 12:24:42.091852 6070 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091859 6070 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 12:24:42.091923 6070 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 12:24:42.091861 6070 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 12:24:42.091944 6070 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 12:24:42.091985 6070 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 12:24:42.092065 6070 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 12:24:42.092107 6070 factory.go:656] Stopping watch factory\\\\nI1007 12:24:42.092128 6070 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 12:24:42.092155 6070 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 12:24:42.092178 6070 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 12:24:42.092192 6070 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.317578 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.329998 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.341217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.341252 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.341262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.341279 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.341289 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.346130 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.360505 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.376197 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.391039 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.405052 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.420721 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.443965 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.444010 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.444019 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.444043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.444054 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.550098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.550140 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.550149 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.550169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.550183 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.652597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.652658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.652675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.652703 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.652723 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.755341 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.755378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.755389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.755407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.755421 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.857769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.857829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.857848 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.857872 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.857917 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.960261 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.960330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.960348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.960372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.960392 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:43Z","lastTransitionTime":"2025-10-07T12:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.962774 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" event={"ID":"ad6b96be-e641-4d3e-a8d8-a750d029fdc5","Type":"ContainerStarted","Data":"b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.962812 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" event={"ID":"ad6b96be-e641-4d3e-a8d8-a750d029fdc5","Type":"ContainerStarted","Data":"d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.962823 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" event={"ID":"ad6b96be-e641-4d3e-a8d8-a750d029fdc5","Type":"ContainerStarted","Data":"40f087d23eae5621af72ae9b6137c172c2ef904e4443ec352e79c0a0ee4cbffb"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.966312 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/0.log" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.971475 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f"} Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.972041 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.982638 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-znndd"] Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.983311 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:43 crc kubenswrapper[4702]: E1007 12:24:43.983398 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:43 crc kubenswrapper[4702]: I1007 12:24:43.988052 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.004911 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.018757 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vczpm\" (UniqueName: \"kubernetes.io/projected/8fb0cea3-d8a9-4738-abf3-47eea11213d7-kube-api-access-vczpm\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.019078 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.025663 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.046161 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.063193 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.063883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.063947 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.063960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.063981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.063995 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.079083 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.102944 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.119585 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.119636 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vczpm\" (UniqueName: \"kubernetes.io/projected/8fb0cea3-d8a9-4738-abf3-47eea11213d7-kube-api-access-vczpm\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.119824 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.119956 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:44.619926855 +0000 UTC m=+40.795663171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.126787 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091824 6070 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 12:24:42.091833 6070 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 12:24:42.091852 6070 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091859 6070 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 12:24:42.091923 6070 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 12:24:42.091861 6070 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 12:24:42.091944 6070 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 12:24:42.091985 6070 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 12:24:42.092065 6070 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 12:24:42.092107 6070 factory.go:656] Stopping watch factory\\\\nI1007 12:24:42.092128 6070 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 12:24:42.092155 6070 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 12:24:42.092178 6070 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 12:24:42.092192 6070 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.142214 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.142554 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vczpm\" (UniqueName: \"kubernetes.io/projected/8fb0cea3-d8a9-4738-abf3-47eea11213d7-kube-api-access-vczpm\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.167663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.167731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.167750 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.167780 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.167809 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.169210 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.192196 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.214944 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.231675 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.249813 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.267507 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.271579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.271645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.271663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.271687 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.271704 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.285721 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.302389 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.322141 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.348166 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.363961 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.374620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.374665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.374684 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.374703 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.374714 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.384527 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.400323 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.416700 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.429611 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.450801 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091824 6070 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 12:24:42.091833 6070 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 12:24:42.091852 6070 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091859 6070 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 12:24:42.091923 6070 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 12:24:42.091861 6070 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 12:24:42.091944 6070 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 12:24:42.091985 6070 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 12:24:42.092065 6070 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 12:24:42.092107 6070 factory.go:656] Stopping watch factory\\\\nI1007 12:24:42.092128 6070 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 12:24:42.092155 6070 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 12:24:42.092178 6070 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 12:24:42.092192 6070 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.462705 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.477561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.477625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.477645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.477676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.477701 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.477773 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.493123 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.511864 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.522583 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.535636 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.582728 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.582844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.582871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.582942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.582972 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.624555 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.624834 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.624997 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:45.624958631 +0000 UTC m=+41.800694787 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.640038 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.640068 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.640258 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.640372 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.640441 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.640606 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.658737 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.676223 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.686400 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.686469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.686480 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.686499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.686511 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.692291 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.710175 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.727238 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.751136 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.766241 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.783179 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.792180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.792450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.792648 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.792868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.793125 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.802511 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.836419 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091824 6070 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 12:24:42.091833 6070 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 12:24:42.091852 6070 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091859 6070 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 12:24:42.091923 6070 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 12:24:42.091861 6070 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 12:24:42.091944 6070 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 12:24:42.091985 6070 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 12:24:42.092065 6070 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 12:24:42.092107 6070 factory.go:656] Stopping watch factory\\\\nI1007 12:24:42.092128 6070 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 12:24:42.092155 6070 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 12:24:42.092178 6070 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 12:24:42.092192 6070 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.851261 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.866163 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.881485 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.893579 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.895308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.895343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.895354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.895374 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.895387 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.905790 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.919761 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.977074 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/1.log" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.977636 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/0.log" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.980102 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f" exitCode=1 Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.980264 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f"} Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.980975 4702 scope.go:117] "RemoveContainer" containerID="084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.981223 4702 scope.go:117] "RemoveContainer" containerID="2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698" Oct 07 12:24:44 crc kubenswrapper[4702]: E1007 12:24:44.981524 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.997760 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.998341 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.998489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.998553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.998624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:44 crc kubenswrapper[4702]: I1007 12:24:44.998693 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:44Z","lastTransitionTime":"2025-10-07T12:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.011489 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.027888 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.042822 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.054504 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.068039 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.089541 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f50a30c07b80b1ab1a448e1b42f76b299d162e8cfc61157398840c23970d698\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091824 6070 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 12:24:42.091833 6070 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 12:24:42.091852 6070 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 12:24:42.091859 6070 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 12:24:42.091923 6070 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 12:24:42.091861 6070 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 12:24:42.091944 6070 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 12:24:42.091985 6070 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 12:24:42.092065 6070 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 12:24:42.092107 6070 factory.go:656] Stopping watch factory\\\\nI1007 12:24:42.092128 6070 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 12:24:42.092155 6070 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 12:24:42.092178 6070 handler.go:208] Removed *v1.Node event handler 7\\\\nI1007 12:24:42.092192 6070 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:44Z\\\",\\\"message\\\":\\\"ons:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.040050 6215 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-route-controller-manager/route-controller-manager]} name:Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.239:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.039378 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 12:24:44.039551 6215 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1007 12:24:44.039094 6215 services_controller.go:453] Built service openshift-marketplace/redhat-marketplace template LB for network=default: []services.LB{}\\\\nF1007 12:24:44.040263 6215 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.102193 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.102291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.102305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.102204 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.102328 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.102504 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.117622 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.136516 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.151287 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.167434 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.180483 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.191311 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.205540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.205587 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.205597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.205620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.205646 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.207745 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.222857 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:45Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.308339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.308399 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.308412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.308437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.308458 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.411310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.411355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.411368 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.411388 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.411401 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.514596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.514714 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.514739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.514769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.514791 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.617961 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.618012 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.618023 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.618047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.618058 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.636611 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:45 crc kubenswrapper[4702]: E1007 12:24:45.636800 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:45 crc kubenswrapper[4702]: E1007 12:24:45.636886 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:47.636862476 +0000 UTC m=+43.812598592 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.640415 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:45 crc kubenswrapper[4702]: E1007 12:24:45.640577 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.722380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.722437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.722449 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.722470 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.722483 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.826118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.826240 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.826275 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.826307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.826329 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.930094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.930177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.930210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.930246 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.930270 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:45Z","lastTransitionTime":"2025-10-07T12:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.986308 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/1.log" Oct 07 12:24:45 crc kubenswrapper[4702]: I1007 12:24:45.990560 4702 scope.go:117] "RemoveContainer" containerID="084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f" Oct 07 12:24:45 crc kubenswrapper[4702]: E1007 12:24:45.990939 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.004963 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.017151 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.029780 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.033124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.033167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.033176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.033195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.033207 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.046165 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.061797 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.080818 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:44Z\\\",\\\"message\\\":\\\"ons:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.040050 6215 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-route-controller-manager/route-controller-manager]} name:Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.239:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.039378 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 12:24:44.039551 6215 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1007 12:24:44.039094 6215 services_controller.go:453] Built service openshift-marketplace/redhat-marketplace template LB for network=default: []services.LB{}\\\\nF1007 12:24:44.040263 6215 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.094678 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.111464 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.129108 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.136635 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.136681 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.136702 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.136730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.136828 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.147951 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.162209 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.177085 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.193411 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.213355 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.230313 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.239985 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.240106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.240138 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.240167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.240185 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.244144 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:46Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.343346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.343391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.343408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.343428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.343441 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.447213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.447301 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.447327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.447370 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.447394 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.550088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.550123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.550133 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.550152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.550163 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.640528 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.640519 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:46 crc kubenswrapper[4702]: E1007 12:24:46.640691 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.640549 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:46 crc kubenswrapper[4702]: E1007 12:24:46.640974 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:46 crc kubenswrapper[4702]: E1007 12:24:46.641170 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.652521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.652575 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.652589 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.652612 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.652627 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.755377 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.755421 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.755431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.755451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.755462 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.858946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.859046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.859066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.859122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.859171 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.962508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.962563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.962574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.962597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:46 crc kubenswrapper[4702]: I1007 12:24:46.962609 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:46Z","lastTransitionTime":"2025-10-07T12:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.066096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.066176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.066197 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.066226 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.066244 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.169361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.169431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.169467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.169502 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.169526 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.272634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.272722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.272757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.272792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.272816 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.376691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.376746 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.376762 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.376787 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.376806 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.480181 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.480253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.480268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.480292 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.480307 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.583942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.583985 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.583996 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.584015 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.584027 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.640526 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:47 crc kubenswrapper[4702]: E1007 12:24:47.640741 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.658191 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:47 crc kubenswrapper[4702]: E1007 12:24:47.658359 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:47 crc kubenswrapper[4702]: E1007 12:24:47.658420 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:51.65840255 +0000 UTC m=+47.834138666 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.687516 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.687579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.687591 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.687618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.687634 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.791319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.791393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.791405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.791427 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.791443 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.894326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.894380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.894389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.894408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.894421 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.997360 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.997431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.997452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.997482 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:47 crc kubenswrapper[4702]: I1007 12:24:47.997498 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:47Z","lastTransitionTime":"2025-10-07T12:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.099922 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.100004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.100025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.100055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.100089 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.203226 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.203263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.203271 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.203288 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.203302 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.305808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.305855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.305865 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.305882 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.305919 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.409047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.409119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.409133 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.409155 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.409169 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.512394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.512442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.512452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.512470 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.512480 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.615362 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.615419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.615433 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.615456 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.615470 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.640991 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.641072 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:48 crc kubenswrapper[4702]: E1007 12:24:48.641175 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.641236 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:48 crc kubenswrapper[4702]: E1007 12:24:48.641391 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:48 crc kubenswrapper[4702]: E1007 12:24:48.641499 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.718614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.718662 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.718675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.718700 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.718713 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.822218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.822728 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.822743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.822762 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.822776 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.925403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.925454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.925470 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.925500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:48 crc kubenswrapper[4702]: I1007 12:24:48.925511 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:48Z","lastTransitionTime":"2025-10-07T12:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.028126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.028195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.028208 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.028229 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.028245 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.131297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.131385 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.131435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.131463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.131482 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.234125 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.234202 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.234218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.234245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.234264 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.337462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.337505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.337514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.337531 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.337545 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.441559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.441600 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.441610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.441628 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.441642 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.544985 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.545038 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.545051 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.545074 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.545085 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.640195 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:49 crc kubenswrapper[4702]: E1007 12:24:49.640383 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.649022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.649071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.649089 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.649115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.649185 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.752766 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.752816 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.752826 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.752844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.752855 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.856885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.856962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.856975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.856997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.857008 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.959849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.959930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.959942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.959962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:49 crc kubenswrapper[4702]: I1007 12:24:49.959975 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:49Z","lastTransitionTime":"2025-10-07T12:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.062494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.062567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.062587 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.062618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.062643 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.166376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.166450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.166465 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.166487 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.166501 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.270591 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.270668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.270685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.270713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.270734 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.374498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.374567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.374584 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.374619 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.374638 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.478739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.478843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.478864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.478893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.478941 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.582796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.582865 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.582883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.582934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.582954 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.640973 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.641010 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.641133 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:50 crc kubenswrapper[4702]: E1007 12:24:50.641357 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:50 crc kubenswrapper[4702]: E1007 12:24:50.641470 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:50 crc kubenswrapper[4702]: E1007 12:24:50.641625 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.686346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.686441 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.686468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.686504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.686536 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.790025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.790096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.790113 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.790143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.790164 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.894178 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.894238 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.894257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.894286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.894328 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.997079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.997177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.997206 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.997240 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:50 crc kubenswrapper[4702]: I1007 12:24:50.997261 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:50Z","lastTransitionTime":"2025-10-07T12:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.100429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.100493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.100507 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.100532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.100549 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.204320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.204375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.204385 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.204408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.204419 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.307241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.307310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.307330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.307364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.307383 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.410475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.410520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.410532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.410557 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.410570 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.512186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.512256 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.512273 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.512299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.512314 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.527687 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:51Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.532361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.532410 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.532428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.532455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.532473 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.562129 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:51Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.568387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.568445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.568462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.568517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.568538 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.597656 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:51Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.614997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.615059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.615137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.615168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.615185 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.633141 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:51Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.637245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.637280 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.637291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.637311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.637323 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.640097 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.640327 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.650629 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:51Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.650758 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.652532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.652592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.652608 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.652630 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.652647 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.713734 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.714025 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:51 crc kubenswrapper[4702]: E1007 12:24:51.714157 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:24:59.714127281 +0000 UTC m=+55.889863407 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.755936 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.756010 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.756031 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.756063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.756084 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.859908 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.859966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.859976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.859996 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.860007 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.963277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.963357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.963377 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.963408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:51 crc kubenswrapper[4702]: I1007 12:24:51.963428 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:51Z","lastTransitionTime":"2025-10-07T12:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.066839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.066934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.066950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.066973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.066987 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.170541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.170614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.170636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.170665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.170684 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.273816 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.273923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.273948 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.273975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.273991 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.377625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.378133 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.378359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.378548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.378737 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.482040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.483190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.483253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.483298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.483329 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.586888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.587018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.587041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.587073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.587099 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.640724 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.641276 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.641353 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:52 crc kubenswrapper[4702]: E1007 12:24:52.641540 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:52 crc kubenswrapper[4702]: E1007 12:24:52.641940 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:52 crc kubenswrapper[4702]: E1007 12:24:52.642155 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.690765 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.690813 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.690831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.690878 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.690940 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.793411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.793777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.793964 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.794142 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.794283 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.897550 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.897613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.897630 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.897657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:52 crc kubenswrapper[4702]: I1007 12:24:52.897676 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:52Z","lastTransitionTime":"2025-10-07T12:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.001408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.001514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.001535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.001562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.001583 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.105280 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.105366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.105391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.105424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.105444 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.208448 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.208521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.208545 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.208576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.208598 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.312191 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.312278 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.312302 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.312340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.312364 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.414979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.415066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.415100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.415131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.415148 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.518754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.518823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.518845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.518877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.519005 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.622178 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.622268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.622293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.622327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.622352 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.639837 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:53 crc kubenswrapper[4702]: E1007 12:24:53.640062 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.725624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.725703 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.725727 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.725760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.725784 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.828770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.828812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.828824 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.828842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.828855 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.932831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.932921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.932941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.932967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:53 crc kubenswrapper[4702]: I1007 12:24:53.932988 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:53Z","lastTransitionTime":"2025-10-07T12:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.035828 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.035889 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.035944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.035994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.036026 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.139818 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.139883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.139932 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.139960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.139980 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.243244 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.243327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.243359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.243392 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.243414 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.346966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.347035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.347055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.347083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.347102 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.450779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.450850 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.450869 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.450921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.450940 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.554783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.554828 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.554838 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.554856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.554868 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.640128 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.640144 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:54 crc kubenswrapper[4702]: E1007 12:24:54.640417 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:54 crc kubenswrapper[4702]: E1007 12:24:54.640548 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.640230 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:54 crc kubenswrapper[4702]: E1007 12:24:54.641378 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.658176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.658254 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.658278 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.658310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.658335 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.662836 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.679528 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.698222 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.715154 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.735751 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.762241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.762299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.762315 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.762338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.762355 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.764334 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:44Z\\\",\\\"message\\\":\\\"ons:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.040050 6215 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-route-controller-manager/route-controller-manager]} name:Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.239:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.039378 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 12:24:44.039551 6215 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1007 12:24:44.039094 6215 services_controller.go:453] Built service openshift-marketplace/redhat-marketplace template LB for network=default: []services.LB{}\\\\nF1007 12:24:44.040263 6215 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.782049 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.806562 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.827980 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.844008 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.857012 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.865811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.865883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.865927 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.865949 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.865962 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.869916 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.883324 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.899187 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.911798 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.925356 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:54Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.969484 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.969794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.969912 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.970082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:54 crc kubenswrapper[4702]: I1007 12:24:54.970156 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:54Z","lastTransitionTime":"2025-10-07T12:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.073341 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.073384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.073392 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.073407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.073418 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.176212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.176261 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.176271 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.176291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.176301 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.279248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.279317 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.279338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.279366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.279385 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.382560 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.382651 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.382675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.382707 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.382728 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.486138 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.486197 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.486217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.486245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.486262 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.588339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.588400 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.588416 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.588444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.588458 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.640387 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:55 crc kubenswrapper[4702]: E1007 12:24:55.640608 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.691583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.691639 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.691656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.691682 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.691705 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.795483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.795553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.795602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.795632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.795652 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.899173 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.899240 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.899266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.899297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:55 crc kubenswrapper[4702]: I1007 12:24:55.899320 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:55Z","lastTransitionTime":"2025-10-07T12:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.003030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.003147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.003183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.003220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.003241 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.107784 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.108335 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.108577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.108789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.108873 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.213331 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.213734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.213862 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.214302 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.214363 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.318291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.318375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.318394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.318426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.318445 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.422266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.422357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.422378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.422404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.422423 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.464942 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.465197 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:25:28.465155353 +0000 UTC m=+84.640891519 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.524876 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.524989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.525016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.525049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.525072 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.566550 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.566634 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.566719 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.566806 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.566877 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.566891 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.566944 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567062 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567107 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567127 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567060 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:25:28.567024902 +0000 UTC m=+84.742761018 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567187 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567216 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567222 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:25:28.567185976 +0000 UTC m=+84.742922262 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567275 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:25:28.567258048 +0000 UTC m=+84.742994414 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.567314 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:25:28.567291269 +0000 UTC m=+84.743027425 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.628444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.628523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.628536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.628556 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.628571 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.640439 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.640616 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.640838 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.640916 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.641084 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:56 crc kubenswrapper[4702]: E1007 12:24:56.641317 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.732354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.732449 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.732476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.732508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.732534 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.836431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.836511 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.836533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.836566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.836590 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.940384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.940466 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.940488 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.940520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:56 crc kubenswrapper[4702]: I1007 12:24:56.940548 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:56Z","lastTransitionTime":"2025-10-07T12:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.043952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.044017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.044034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.044057 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.044074 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.147545 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.147650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.147675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.147711 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.147736 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.250735 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.250786 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.250796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.250815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.250826 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.355136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.355200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.355222 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.355248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.355268 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.458769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.458850 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.458876 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.458943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.458982 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.561761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.561815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.561827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.561853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.561868 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.640693 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:57 crc kubenswrapper[4702]: E1007 12:24:57.640877 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.664858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.664922 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.664935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.664957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.664970 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.768268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.768319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.768330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.768355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.768367 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.871356 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.871411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.871424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.871446 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.871463 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.975159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.975250 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.975269 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.975298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:57 crc kubenswrapper[4702]: I1007 12:24:57.975318 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:57Z","lastTransitionTime":"2025-10-07T12:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.078497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.078564 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.078579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.078630 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.078648 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.160271 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.171642 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.179393 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.182072 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.182137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.182158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.182186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.182205 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.195800 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.216834 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.236486 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.249720 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.264603 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.285561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.285657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.285687 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.285716 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.285741 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.292464 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:44Z\\\",\\\"message\\\":\\\"ons:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.040050 6215 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-route-controller-manager/route-controller-manager]} name:Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.239:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.039378 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 12:24:44.039551 6215 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1007 12:24:44.039094 6215 services_controller.go:453] Built service openshift-marketplace/redhat-marketplace template LB for network=default: []services.LB{}\\\\nF1007 12:24:44.040263 6215 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.313129 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.334310 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.354632 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.376779 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.389404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.389474 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.389494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.389531 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.389572 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.403307 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.421028 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.438877 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.453631 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.468378 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:24:58Z is after 2025-08-24T17:21:41Z" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.493099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.493161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.493175 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.493197 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.493212 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.599975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.600077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.600093 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.600114 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.600131 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.640317 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.640351 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:24:58 crc kubenswrapper[4702]: E1007 12:24:58.640552 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:24:58 crc kubenswrapper[4702]: E1007 12:24:58.640695 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.640761 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:24:58 crc kubenswrapper[4702]: E1007 12:24:58.640955 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.703748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.703790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.703800 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.703818 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.703829 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.807113 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.807171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.807182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.807206 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.807222 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.910299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.910406 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.910434 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.910472 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:58 crc kubenswrapper[4702]: I1007 12:24:58.910499 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:58Z","lastTransitionTime":"2025-10-07T12:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.013131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.013210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.013235 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.013267 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.013293 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.115873 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.116403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.116543 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.116691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.116849 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.220384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.220455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.220477 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.220507 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.220531 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.323872 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.323976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.324000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.324029 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.324056 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.427095 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.427155 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.427164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.427182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.427193 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.530974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.531055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.531069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.531094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.531110 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.633978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.634025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.634034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.634051 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.634063 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.640525 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:59 crc kubenswrapper[4702]: E1007 12:24:59.640715 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.737443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.737536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.737561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.737597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.737620 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.810870 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:24:59 crc kubenswrapper[4702]: E1007 12:24:59.811152 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:59 crc kubenswrapper[4702]: E1007 12:24:59.811282 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:25:15.811256104 +0000 UTC m=+71.986992440 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.841119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.841171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.841186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.841205 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.841216 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.944030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.944122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.944152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.944187 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:24:59 crc kubenswrapper[4702]: I1007 12:24:59.944214 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:24:59Z","lastTransitionTime":"2025-10-07T12:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.046849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.046955 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.046974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.046999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.047018 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.150731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.150782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.150792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.150812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.150823 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.255205 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.255258 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.255281 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.255309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.255331 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.359663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.359722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.359739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.359771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.359786 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.463232 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.463297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.463315 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.463344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.463362 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.567938 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.568048 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.568075 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.568104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.568126 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.640389 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.640476 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:00 crc kubenswrapper[4702]: E1007 12:25:00.640546 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:00 crc kubenswrapper[4702]: E1007 12:25:00.640644 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.640677 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:00 crc kubenswrapper[4702]: E1007 12:25:00.640788 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.641375 4702 scope.go:117] "RemoveContainer" containerID="084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.671642 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.672056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.672071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.672088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.672100 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.774481 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.774540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.774558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.774576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.774589 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.877401 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.877453 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.877461 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.877479 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.877492 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.980137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.980183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.980195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.980213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:00 crc kubenswrapper[4702]: I1007 12:25:00.980226 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:00Z","lastTransitionTime":"2025-10-07T12:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.052391 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/1.log" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.055434 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.056034 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.076929 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.083048 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.083109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.083127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.083146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.083160 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.093148 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.109058 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.123377 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.144702 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.164591 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.176722 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.185923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.185965 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.185977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.186012 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.186028 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.189758 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.201793 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.217704 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.237579 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:44Z\\\",\\\"message\\\":\\\"ons:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.040050 6215 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-route-controller-manager/route-controller-manager]} name:Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.239:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.039378 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 12:24:44.039551 6215 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1007 12:24:44.039094 6215 services_controller.go:453] Built service openshift-marketplace/redhat-marketplace template LB for network=default: []services.LB{}\\\\nF1007 12:24:44.040263 6215 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.252654 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.273387 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.285532 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.288478 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.288534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.288547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.288567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.288584 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.301672 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.313044 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.325422 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.391056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.391125 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.391143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.391164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.391178 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.494728 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.494782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.494794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.494817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.494832 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.598193 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.598245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.598262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.598291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.598311 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.639859 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.640099 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.690347 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.690423 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.690437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.690457 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.690472 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.707635 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.712087 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.712127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.712138 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.712161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.712177 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.728357 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.734605 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.734652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.734670 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.734693 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.734709 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.755112 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.761119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.761182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.761201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.761229 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.761251 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.781839 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.788412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.788465 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.788492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.788526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.788555 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.808977 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:01 crc kubenswrapper[4702]: E1007 12:25:01.809157 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.812738 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.812820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.812848 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.812892 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.812959 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.917445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.917517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.917539 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.917566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:01 crc kubenswrapper[4702]: I1007 12:25:01.917585 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:01Z","lastTransitionTime":"2025-10-07T12:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.020251 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.020323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.020345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.020375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.020396 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.061337 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/2.log" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.062426 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/1.log" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.066316 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d" exitCode=1 Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.066378 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.066447 4702 scope.go:117] "RemoveContainer" containerID="084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.067240 4702 scope.go:117] "RemoveContainer" containerID="367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d" Oct 07 12:25:02 crc kubenswrapper[4702]: E1007 12:25:02.067452 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.091632 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.118482 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.124324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.124380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.124393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.124414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.124801 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.138330 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.162082 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.180140 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.194731 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.211547 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.227434 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.227483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.227495 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.227514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.227528 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.228562 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.242359 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.254406 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.270628 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.286677 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.301200 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.315558 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.330415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.330477 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.330498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.330525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.330540 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.335163 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.357945 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://084ca55276ea56c549718849d4de19802acea0a82496059cca037d5de47b292f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:24:44Z\\\",\\\"message\\\":\\\"ons:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.040050 6215 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-route-controller-manager/route-controller-manager]} name:Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.239:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {18746a4d-8a63-458a-b7e3-8fb89ff95fc0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 12:24:44.039378 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 12:24:44.039551 6215 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1007 12:24:44.039094 6215 services_controller.go:453] Built service openshift-marketplace/redhat-marketplace template LB for network=default: []services.LB{}\\\\nF1007 12:24:44.040263 6215 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.374103 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:02Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.433343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.433407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.433425 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.433455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.433477 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.536874 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.536954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.536968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.536990 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.537008 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640003 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640056 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640070 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:02 crc kubenswrapper[4702]: E1007 12:25:02.640146 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640453 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: E1007 12:25:02.640461 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:02 crc kubenswrapper[4702]: E1007 12:25:02.640356 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640478 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.640545 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.743939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.743997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.744017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.744044 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.744068 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.847560 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.847663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.847698 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.847729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.847753 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.951127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.951199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.951224 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.951259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:02 crc kubenswrapper[4702]: I1007 12:25:02.951287 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:02Z","lastTransitionTime":"2025-10-07T12:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.055055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.055134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.055158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.055192 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.055215 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.072873 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/2.log" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.077842 4702 scope.go:117] "RemoveContainer" containerID="367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d" Oct 07 12:25:03 crc kubenswrapper[4702]: E1007 12:25:03.078423 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.096355 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.114176 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.133522 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.153678 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.158316 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.158386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.158412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.158442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.158464 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.172570 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.192488 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.214228 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.234740 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.256740 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.261227 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.261295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.261315 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.261343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.261359 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.277236 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.296950 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.323463 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.341150 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.360890 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.364000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.364042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.364053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.364070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.364082 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.376749 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.391137 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.404010 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:03Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.467659 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.467737 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.467762 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.467793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.467876 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.572102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.572164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.572179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.572204 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.572220 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.640915 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:03 crc kubenswrapper[4702]: E1007 12:25:03.641107 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.676254 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.676319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.676338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.676362 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.676376 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.779661 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.779704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.779725 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.779748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.779763 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.883333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.883409 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.883432 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.883463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.883488 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.987664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.987789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.987814 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.987846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:03 crc kubenswrapper[4702]: I1007 12:25:03.987871 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:03Z","lastTransitionTime":"2025-10-07T12:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.091195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.091252 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.091263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.091284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.091299 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.194369 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.194447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.194463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.194491 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.194510 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.298346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.298423 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.298438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.298464 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.298489 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.402111 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.402168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.402183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.402204 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.402220 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.505954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.506030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.506053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.506084 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.506108 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.610238 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.610335 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.610354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.610380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.610399 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.640622 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.640707 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.640629 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:04 crc kubenswrapper[4702]: E1007 12:25:04.640868 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:04 crc kubenswrapper[4702]: E1007 12:25:04.641127 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:04 crc kubenswrapper[4702]: E1007 12:25:04.641229 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.664938 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.683419 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.703761 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.713675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.714193 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.714221 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.714253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.714277 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.720858 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.741369 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.761933 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.775462 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.791719 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.807028 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.820606 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.820869 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.820885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.820932 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.820948 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.829754 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.845140 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.856295 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.865502 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.878208 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.894710 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.914977 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.924265 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.924352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.924376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.924399 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.924437 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:04Z","lastTransitionTime":"2025-10-07T12:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:04 crc kubenswrapper[4702]: I1007 12:25:04.931950 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:04Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.027259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.027326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.027338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.027358 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.027375 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.131106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.131183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.131201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.131236 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.131259 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.234873 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.234940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.234954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.234973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.234988 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.338588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.338800 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.338890 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.338999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.339078 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.443377 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.443468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.443488 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.443518 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.443537 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.547562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.547641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.547671 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.547722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.547753 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.640646 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:05 crc kubenswrapper[4702]: E1007 12:25:05.640834 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.650291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.650330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.650339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.650355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.650366 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.753561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.753599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.753610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.753633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.753644 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.856997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.857064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.857084 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.857114 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.857137 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.960753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.960808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.960820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.960842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:05 crc kubenswrapper[4702]: I1007 12:25:05.960856 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:05Z","lastTransitionTime":"2025-10-07T12:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.064785 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.064860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.064880 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.064944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.064964 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.168212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.168303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.168326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.168360 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.168385 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.271954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.272035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.272063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.272094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.272117 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.375122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.375183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.375200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.375218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.375230 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.478272 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.478324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.478340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.478358 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.478370 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.581793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.581860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.581872 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.581919 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.581934 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.639857 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.639984 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.640008 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:06 crc kubenswrapper[4702]: E1007 12:25:06.640061 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:06 crc kubenswrapper[4702]: E1007 12:25:06.640140 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:06 crc kubenswrapper[4702]: E1007 12:25:06.640883 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.685793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.685881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.685924 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.685955 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.685975 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.790311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.790371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.790389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.790414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.790433 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.894665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.894750 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.894767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.894795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.894818 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.998699 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.998771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.998805 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.998835 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:06 crc kubenswrapper[4702]: I1007 12:25:06.998861 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:06Z","lastTransitionTime":"2025-10-07T12:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.102334 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.102399 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.102419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.102449 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.102469 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.205183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.205258 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.205277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.205305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.205326 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.309188 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.309245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.309263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.309293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.309329 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.412012 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.412070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.412088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.412117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.412137 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.515985 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.516033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.516045 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.516069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.516083 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.618326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.618390 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.618402 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.618426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.618449 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.639974 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:07 crc kubenswrapper[4702]: E1007 12:25:07.640158 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.721436 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.721505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.721521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.721546 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.721567 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.825259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.825318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.825329 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.825349 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.825362 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.928404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.928471 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.928493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.928517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:07 crc kubenswrapper[4702]: I1007 12:25:07.928535 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:07Z","lastTransitionTime":"2025-10-07T12:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.031351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.031393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.031405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.031424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.031435 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.134646 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.134732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.134748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.134780 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.134801 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.238183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.238302 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.238321 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.238347 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.238369 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.341330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.341375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.341389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.341408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.341423 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.444802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.444997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.445025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.445062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.445087 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.548582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.548643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.548657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.548677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.548696 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.640623 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.640752 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.640814 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:08 crc kubenswrapper[4702]: E1007 12:25:08.641093 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:08 crc kubenswrapper[4702]: E1007 12:25:08.641261 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:08 crc kubenswrapper[4702]: E1007 12:25:08.641455 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.652191 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.652270 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.652309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.652347 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.652376 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.755682 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.755794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.755814 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.755844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.755865 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.859702 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.859770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.859790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.859815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.859835 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.963731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.963789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.963805 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.963829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:08 crc kubenswrapper[4702]: I1007 12:25:08.963850 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:08Z","lastTransitionTime":"2025-10-07T12:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.068022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.068108 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.068132 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.068168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.068192 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.171831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.171880 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.171926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.171953 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.171972 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.287016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.287097 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.287128 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.287160 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.287181 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.391405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.391458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.391475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.391498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.391518 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.495027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.495077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.495094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.495122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.495141 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.598839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.598934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.598953 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.598978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.598996 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.640656 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:09 crc kubenswrapper[4702]: E1007 12:25:09.640809 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.702563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.702628 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.702656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.702686 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.702708 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.806863 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.806980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.807008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.807040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.807065 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.910645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.910743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.910802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.910838 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:09 crc kubenswrapper[4702]: I1007 12:25:09.910863 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:09Z","lastTransitionTime":"2025-10-07T12:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.015117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.015177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.015196 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.015224 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.015244 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.119130 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.119190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.119203 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.119226 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.119240 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.222297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.222370 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.222392 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.222422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.222443 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.326297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.326398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.326415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.326443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.326461 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.429946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.429995 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.430012 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.430037 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.430056 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.533614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.533744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.533763 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.533796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.533813 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.637770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.637857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.637871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.637913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.637932 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.640305 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.640398 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.640447 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:10 crc kubenswrapper[4702]: E1007 12:25:10.640464 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:10 crc kubenswrapper[4702]: E1007 12:25:10.640549 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:10 crc kubenswrapper[4702]: E1007 12:25:10.640651 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.742308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.742423 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.742440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.742464 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.742479 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.845918 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.845957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.845967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.845984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.845995 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.949521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.949577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.949592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.949615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:10 crc kubenswrapper[4702]: I1007 12:25:10.949629 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:10Z","lastTransitionTime":"2025-10-07T12:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.052811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.052857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.052875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.052910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.052926 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.155796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.155829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.155839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.155853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.155863 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.258263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.258289 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.258298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.258311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.258319 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.361319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.361379 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.361395 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.361422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.361440 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.465082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.465157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.465183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.465255 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.465279 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.567986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.568062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.568098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.568140 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.568168 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.640636 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.640820 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.671216 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.671272 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.671284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.671306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.671318 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.774589 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.774650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.774664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.774688 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.774700 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.877206 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.877263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.877277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.877304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.877320 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.917120 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.917170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.917184 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.917204 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.917215 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.932612 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:11Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.937308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.937382 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.937403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.937434 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.937455 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.951232 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:11Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.955296 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.955350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.955363 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.955383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.955396 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.968053 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:11Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.972741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.972779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.972788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.972806 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.972817 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.984503 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:11Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.988092 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.988118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.988128 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.988145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:11 crc kubenswrapper[4702]: I1007 12:25:11.988158 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:11Z","lastTransitionTime":"2025-10-07T12:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.999147 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:11Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:11 crc kubenswrapper[4702]: E1007 12:25:11.999282 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.000722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.000747 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.000757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.000769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.000779 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.103856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.103884 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.103910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.103928 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.103940 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.207006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.207053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.207070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.207087 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.207097 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.309758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.309821 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.309834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.309875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.309923 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.413278 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.413339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.413360 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.413386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.413405 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.516121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.516161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.516170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.516186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.516200 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.619371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.619445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.619460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.619482 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.619500 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.640839 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.641029 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.641416 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:12 crc kubenswrapper[4702]: E1007 12:25:12.641626 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:12 crc kubenswrapper[4702]: E1007 12:25:12.641711 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:12 crc kubenswrapper[4702]: E1007 12:25:12.641805 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.651825 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.722614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.722657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.722668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.722689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.722703 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.825632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.825663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.825673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.825692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.825702 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.928387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.928455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.928474 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.928492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:12 crc kubenswrapper[4702]: I1007 12:25:12.928522 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:12Z","lastTransitionTime":"2025-10-07T12:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.030920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.030981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.030990 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.031009 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.031019 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.133959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.134028 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.134048 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.134078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.134096 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.237212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.238097 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.238454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.238678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.238775 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.342788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.343385 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.343636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.343852 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.344148 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.447063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.447129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.447143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.447166 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.447181 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.549684 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.549746 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.549758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.549782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.549796 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.640697 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:13 crc kubenswrapper[4702]: E1007 12:25:13.641277 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.653027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.653066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.653078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.653100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.653113 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.756384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.756424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.756433 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.756451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.756461 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.860446 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.860505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.860518 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.860543 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.860555 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.964542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.964582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.964594 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.964610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:13 crc kubenswrapper[4702]: I1007 12:25:13.964622 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:13Z","lastTransitionTime":"2025-10-07T12:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.068442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.068506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.068520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.068540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.068554 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.171734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.172179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.172284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.172361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.172431 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.276266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.276328 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.276344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.276362 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.276375 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.380172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.380247 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.380265 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.380290 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.380308 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.483196 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.483248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.483259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.483277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.483289 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.586788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.586834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.586846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.586866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.586879 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.640049 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.640173 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:14 crc kubenswrapper[4702]: E1007 12:25:14.640220 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.640275 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:14 crc kubenswrapper[4702]: E1007 12:25:14.640448 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:14 crc kubenswrapper[4702]: E1007 12:25:14.640649 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.658847 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.674603 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.688976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.689018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.689030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.689047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.689059 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.689314 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.701415 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.717849 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.742938 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.757177 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.775201 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.790003 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.791653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.791713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.791759 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.791801 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.791846 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.803818 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.819956 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.832191 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.844857 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.856311 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.868595 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.892095 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.894930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.894973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.894985 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.895009 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.895025 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.917528 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.936039 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:14Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.996934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.997209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.997305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.997393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:14 crc kubenswrapper[4702]: I1007 12:25:14.997484 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:14Z","lastTransitionTime":"2025-10-07T12:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.099910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.100273 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.100355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.100428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.100488 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.203216 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.203287 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.203307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.203335 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.203354 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.305773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.305812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.305823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.305842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.305857 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.408958 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.408990 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.408999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.409017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.409025 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.512372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.512424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.512435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.512457 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.512472 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.615781 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.615873 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.615980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.616022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.616082 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.640155 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:15 crc kubenswrapper[4702]: E1007 12:25:15.640319 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.719606 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.719653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.719664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.719685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.719700 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.822494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.822548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.822592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.822611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.822620 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.902169 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:15 crc kubenswrapper[4702]: E1007 12:25:15.902375 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:25:15 crc kubenswrapper[4702]: E1007 12:25:15.902496 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:25:47.902462984 +0000 UTC m=+104.078199140 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.925637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.925690 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.925707 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.925736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:15 crc kubenswrapper[4702]: I1007 12:25:15.925755 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:15Z","lastTransitionTime":"2025-10-07T12:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.028789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.028841 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.028852 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.028869 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.028880 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.130758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.130815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.130825 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.130847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.130856 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.233949 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.233998 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.234012 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.234033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.234047 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.337248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.337350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.337415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.337448 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.337508 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.441641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.441712 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.441731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.441761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.441783 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.545815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.545860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.545868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.545885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.545907 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.639853 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.640006 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.640020 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:16 crc kubenswrapper[4702]: E1007 12:25:16.640125 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:16 crc kubenswrapper[4702]: E1007 12:25:16.640482 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:16 crc kubenswrapper[4702]: E1007 12:25:16.640591 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.640861 4702 scope.go:117] "RemoveContainer" containerID="367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d" Oct 07 12:25:16 crc kubenswrapper[4702]: E1007 12:25:16.641078 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.648449 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.648486 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.648500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.648517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.648528 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.751220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.751280 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.751293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.751313 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.751327 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.854275 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.854324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.854342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.854366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.854379 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.957212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.957271 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.957282 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.957301 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:16 crc kubenswrapper[4702]: I1007 12:25:16.957313 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:16Z","lastTransitionTime":"2025-10-07T12:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.060624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.060677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.060689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.060711 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.060724 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.164775 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.164841 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.164860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.164881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.164927 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.268940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.268978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.268989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.269005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.269015 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.371951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.372001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.372011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.372032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.372042 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.475174 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.475232 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.475243 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.475263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.475274 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.578374 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.578430 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.578443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.578467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.578482 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.640196 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:17 crc kubenswrapper[4702]: E1007 12:25:17.640721 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.681520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.681568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.681582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.681607 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.681625 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.784311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.784394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.784413 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.784438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.784458 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.887847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.887993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.888016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.888040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.888059 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.991367 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.991424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.991444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.991468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:17 crc kubenswrapper[4702]: I1007 12:25:17.991482 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:17Z","lastTransitionTime":"2025-10-07T12:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.094214 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.094268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.094285 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.094306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.094319 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.197831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.197938 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.197954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.197978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.197994 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.300923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.300974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.300987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.301013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.301026 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.402761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.402806 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.402816 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.402833 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.402844 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.505554 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.505628 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.505648 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.505678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.505697 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.609090 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.609159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.609177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.609209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.609228 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.640793 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.640855 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:18 crc kubenswrapper[4702]: E1007 12:25:18.641014 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.641067 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:18 crc kubenswrapper[4702]: E1007 12:25:18.641148 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:18 crc kubenswrapper[4702]: E1007 12:25:18.641243 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.712439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.712502 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.712516 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.712536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.712549 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.815468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.815517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.815528 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.815547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.815562 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.919320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.919411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.919437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.919469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:18 crc kubenswrapper[4702]: I1007 12:25:18.919493 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:18Z","lastTransitionTime":"2025-10-07T12:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.022334 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.022407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.022423 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.022445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.022462 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.125348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.125704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.125715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.125733 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.125745 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.137296 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/0.log" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.137338 4702 generic.go:334] "Generic (PLEG): container finished" podID="943865b8-ec83-4c7e-9e71-b02eaea71c1d" containerID="8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe" exitCode=1 Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.137367 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerDied","Data":"8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.138146 4702 scope.go:117] "RemoveContainer" containerID="8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.175561 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.195864 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.210948 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.229295 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.232529 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.232577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.232588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.232610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.232622 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.244980 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.256842 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.270322 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.286783 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.302198 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.322776 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.335557 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.335616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.335631 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.335673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.335694 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.337612 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.354287 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.372017 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.390234 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.406748 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.420854 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.438030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.438066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.438075 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.438093 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.438104 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.443199 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.454869 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:19Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.541376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.541467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.541476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.541492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.541505 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.640268 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:19 crc kubenswrapper[4702]: E1007 12:25:19.640546 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.644535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.644605 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.644619 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.644641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.644656 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.747105 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.747145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.747156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.747173 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.747184 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.850008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.850049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.850059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.850076 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.850088 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.953532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.953587 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.953602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.953623 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:19 crc kubenswrapper[4702]: I1007 12:25:19.953637 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:19Z","lastTransitionTime":"2025-10-07T12:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.056268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.056311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.056323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.056343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.056355 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.143819 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/0.log" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.143882 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerStarted","Data":"c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.159004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.159032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.159043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.159063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.159083 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.163785 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.186215 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.205252 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.220725 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.237410 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.251099 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.262845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.262890 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.262931 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.262952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.262969 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.266916 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.283678 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.296567 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.312948 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.327290 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.343556 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.369056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.369093 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.369103 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.369121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.369131 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.371302 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.387200 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.401017 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.418550 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.432061 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.445657 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:20Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.471603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.471637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.471646 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.471663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.471674 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.574987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.575025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.575034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.575051 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.575060 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.640221 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.640318 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.640240 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:20 crc kubenswrapper[4702]: E1007 12:25:20.640472 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:20 crc kubenswrapper[4702]: E1007 12:25:20.640583 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:20 crc kubenswrapper[4702]: E1007 12:25:20.640692 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.677662 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.677709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.677722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.677741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.677753 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.780219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.780298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.780321 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.780351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.780374 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.882779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.882831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.882846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.882866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.882882 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.985984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.986028 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.986040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.986057 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:20 crc kubenswrapper[4702]: I1007 12:25:20.986069 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:20Z","lastTransitionTime":"2025-10-07T12:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.089162 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.089209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.089220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.089241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.089257 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.191604 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.191680 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.191700 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.191729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.191749 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.294554 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.294585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.294593 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.294611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.294620 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.398308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.398355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.398371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.398398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.398418 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.501100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.501213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.501232 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.501264 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.501288 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.604175 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.604220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.604233 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.604255 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.604267 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.640723 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:21 crc kubenswrapper[4702]: E1007 12:25:21.640875 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.707124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.707172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.707185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.707211 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.707226 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.810668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.810736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.810754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.810782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.810798 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.913817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.913879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.913892 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.913941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:21 crc kubenswrapper[4702]: I1007 12:25:21.913957 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:21Z","lastTransitionTime":"2025-10-07T12:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.017615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.017685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.017695 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.017717 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.017729 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.121435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.121506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.121520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.121546 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.121560 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.225100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.225173 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.225200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.225236 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.225262 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.329238 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.329303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.329316 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.329339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.329352 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.399143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.399190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.399200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.399219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.399231 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.418542 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:22Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.429382 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.429499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.429874 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.430234 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.430309 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.454348 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:22Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.461289 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.461635 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.462094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.462509 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.462866 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.485807 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:22Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.491201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.491256 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.491273 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.491297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.491314 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.506482 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:22Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.511885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.511950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.511962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.511981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.512024 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.530675 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:22Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.530806 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.533095 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.533153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.533170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.533195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.533210 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.636719 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.637182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.637339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.637507 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.637676 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.640515 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.640518 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.640662 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.640748 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.640957 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:22 crc kubenswrapper[4702]: E1007 12:25:22.640997 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.742101 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.742458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.742523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.742598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.742700 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.845723 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.845781 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.845799 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.845826 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.845845 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.949838 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.949942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.949958 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.949982 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:22 crc kubenswrapper[4702]: I1007 12:25:22.949997 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:22Z","lastTransitionTime":"2025-10-07T12:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.053167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.053249 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.053261 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.053281 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.053292 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.155330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.155380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.155406 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.155426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.155440 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.258410 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.258450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.258475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.258501 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.258520 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.361692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.361748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.361758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.361777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.361794 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.464676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.464730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.464744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.464764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.464780 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.568437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.568532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.568552 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.568581 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.568604 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.640968 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:23 crc kubenswrapper[4702]: E1007 12:25:23.641212 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.671268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.671348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.671370 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.671397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.671417 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.775446 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.775553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.775575 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.775609 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.775631 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.880067 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.880157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.880185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.880217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.880242 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.983496 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.983551 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.983569 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.983590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:23 crc kubenswrapper[4702]: I1007 12:25:23.983601 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:23Z","lastTransitionTime":"2025-10-07T12:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.087231 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.087340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.087357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.087383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.087403 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.190805 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.190893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.190966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.191014 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.191039 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.296225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.296282 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.296299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.296328 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.296345 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.399303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.399338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.399349 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.399368 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.399382 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.503208 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.503262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.503273 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.503294 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.503307 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.606561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.606610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.606622 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.606640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.606650 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.640429 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.640469 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:24 crc kubenswrapper[4702]: E1007 12:25:24.640606 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:24 crc kubenswrapper[4702]: E1007 12:25:24.640725 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.640755 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:24 crc kubenswrapper[4702]: E1007 12:25:24.640950 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.662258 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.679024 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.695286 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.709735 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.709795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.709815 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.709839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.709857 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.715953 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.760720 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.795433 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.811835 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.811879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.811909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.811933 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.811948 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.812380 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.832916 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.852634 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.867054 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.883199 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.903708 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.914346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.914403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.914418 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.914443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.914457 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:24Z","lastTransitionTime":"2025-10-07T12:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.917957 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.935143 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.950728 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.967843 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:24 crc kubenswrapper[4702]: I1007 12:25:24.988801 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:24Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.004402 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:25Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.016925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.016993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.017008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.017032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.017048 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.118927 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.118977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.118989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.119008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.119020 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.221674 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.221735 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.221744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.221763 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.221792 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.325079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.325136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.325151 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.325176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.325192 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.428837 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.428927 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.428946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.428976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.428991 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.532722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.532794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.532808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.532827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.532838 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.636093 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.636165 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.636183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.636203 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.636216 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.640582 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:25 crc kubenswrapper[4702]: E1007 12:25:25.640795 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.740795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.740888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.740962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.740998 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.741023 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.844533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.844580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.844592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.844613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.844626 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.947736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.947808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.947826 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.947852 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:25 crc kubenswrapper[4702]: I1007 12:25:25.947871 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:25Z","lastTransitionTime":"2025-10-07T12:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.050793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.050838 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.050847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.050866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.050876 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.154364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.154411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.154422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.154440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.154454 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.257219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.257269 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.257286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.257316 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.257335 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.360549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.360593 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.360604 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.360628 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.360642 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.463292 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.463364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.463375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.463397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.463408 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.566458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.566522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.566540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.566566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.566584 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.640801 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.640867 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.640955 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:26 crc kubenswrapper[4702]: E1007 12:25:26.641128 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:26 crc kubenswrapper[4702]: E1007 12:25:26.641316 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:26 crc kubenswrapper[4702]: E1007 12:25:26.641586 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.669634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.669690 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.669709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.669734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.669753 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.772760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.772828 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.772845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.772871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.772889 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.876443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.876524 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.876559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.876621 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.876644 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.980544 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.980598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.980608 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.980627 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:26 crc kubenswrapper[4702]: I1007 12:25:26.980638 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:26Z","lastTransitionTime":"2025-10-07T12:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.084043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.084113 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.084129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.084164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.084182 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.187535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.187603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.187616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.187640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.187656 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.292345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.292408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.292425 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.292450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.292468 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.396684 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.396793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.396816 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.396844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.396864 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.500364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.500445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.500465 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.500493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.500516 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.604867 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.604949 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.604968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.604995 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.605023 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.639933 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:27 crc kubenswrapper[4702]: E1007 12:25:27.640108 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.708153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.708207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.708220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.708240 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.708251 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.811438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.811499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.811514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.811538 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.811558 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.918978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.919090 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.919112 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.919143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:27 crc kubenswrapper[4702]: I1007 12:25:27.919165 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:27Z","lastTransitionTime":"2025-10-07T12:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.023063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.023151 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.023169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.023199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.023221 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.125824 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.125889 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.125933 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.125960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.125977 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.230033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.230088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.230102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.230124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.230140 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.334022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.334079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.334099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.334121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.334144 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.437424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.437525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.437552 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.437588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.437612 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.541673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.541751 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.541768 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.541799 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.541818 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.543174 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.543489 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.54343757 +0000 UTC m=+148.719173726 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.640645 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.640702 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.640871 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.640703 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.641028 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.641285 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.642775 4702 scope.go:117] "RemoveContainer" containerID="367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.645726 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.645774 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.645787 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.645807 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.645822 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.646235 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.646401 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.646474 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646507 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646549 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646576 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.646511 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646655 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646848 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646668 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.64663447 +0000 UTC m=+148.822370626 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646890 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646666 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.647023 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.64698753 +0000 UTC m=+148.822723676 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.647094 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.647041662 +0000 UTC m=+148.822777968 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.646693 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: E1007 12:25:28.647189 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.647169955 +0000 UTC m=+148.822906111 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.750776 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.750847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.750865 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.750893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.750944 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.853930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.854387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.854409 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.854435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.854453 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.957370 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.957407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.957420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.957438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:28 crc kubenswrapper[4702]: I1007 12:25:28.957448 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:28Z","lastTransitionTime":"2025-10-07T12:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.060430 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.060463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.060473 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.060490 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.060502 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.162664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.162705 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.162717 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.162738 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.162749 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.178101 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/2.log" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.181046 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.181447 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.198618 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.231719 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.251516 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.264549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.264590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.264605 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.264624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.264636 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.266399 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.292333 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.305293 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.318837 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.331127 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.343853 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.363025 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.369011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.369054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.369088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.369110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.369120 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.378659 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.395417 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.407250 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.419873 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.439452 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.460218 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.472564 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.472635 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.472650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.472676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.472691 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.476954 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.492979 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:29Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.575607 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.575643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.575651 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.575669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.575678 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.640162 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:29 crc kubenswrapper[4702]: E1007 12:25:29.640367 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.679676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.679769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.679814 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.679848 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.679872 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.783751 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.783814 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.783831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.783856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.783873 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.887773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.887837 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.887855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.887883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.887941 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.990822 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.990889 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.990947 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.990983 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:29 crc kubenswrapper[4702]: I1007 12:25:29.991007 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:29Z","lastTransitionTime":"2025-10-07T12:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.094376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.094484 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.094505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.094535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.094554 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.188358 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/3.log" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.189548 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/2.log" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.193290 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" exitCode=1 Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.193344 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.193395 4702 scope.go:117] "RemoveContainer" containerID="367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.194552 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:25:30 crc kubenswrapper[4702]: E1007 12:25:30.194874 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.197100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.197270 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.197304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.197384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.197415 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.214998 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.234020 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.257568 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://367c41434789a844d020d0d561cb09806804c4cf628f56a17e98a278f629f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:01Z\\\",\\\"message\\\":\\\"Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1007 12:25:01.506152 6434 services_controller.go:444] Built service openshift-config-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1007 12:25:01.506178 6434 services_controller.go:445] Built service openshift-config-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1007 12:25:01.506155 6434 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:01Z is after 2025-08-24T17:21:41Z]\\\\nI1007 12:25:01.506208 6434 services_controller.go:451] Built service openshift-conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:29Z\\\",\\\"message\\\":\\\"anager/kube-controller-manager-crc openshift-multus/network-metrics-daemon-znndd openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-multus/multus-additional-cni-plugins-t9vtm openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1007 12:25:29.542730 6817 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1007 12:25:29.542760 6817 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542778 6817 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542791 6817 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1007 12:25:29.542806 6817 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nF1007 12:25:29.542812 6817 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.274300 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.291776 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.300403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.300482 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.300521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.300547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.300559 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.309145 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.323749 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.339884 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.364416 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.382220 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.401092 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.403017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.403052 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.403064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.403088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.403100 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.422027 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.440538 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.457297 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.474238 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.489055 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.503758 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.505523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.505551 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.505561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.505582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.505593 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.519030 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:30Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.608573 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.608621 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.608632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.608655 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.608667 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.640403 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.640468 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.640403 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:30 crc kubenswrapper[4702]: E1007 12:25:30.641076 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:30 crc kubenswrapper[4702]: E1007 12:25:30.641613 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:30 crc kubenswrapper[4702]: E1007 12:25:30.643342 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.712439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.712519 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.712538 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.712572 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.712594 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.816372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.816428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.816445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.816469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.816487 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.919420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.919503 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.919521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.919546 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:30 crc kubenswrapper[4702]: I1007 12:25:30.919569 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:30Z","lastTransitionTime":"2025-10-07T12:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.026318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.026393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.026411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.026439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.026455 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.129614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.129665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.129679 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.129697 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.129712 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.200260 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/3.log" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.205027 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:25:31 crc kubenswrapper[4702]: E1007 12:25:31.205214 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.223718 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.233673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.233743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.233757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.233780 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.233797 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.240672 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.261793 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.283814 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:29Z\\\",\\\"message\\\":\\\"anager/kube-controller-manager-crc openshift-multus/network-metrics-daemon-znndd openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-multus/multus-additional-cni-plugins-t9vtm openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1007 12:25:29.542730 6817 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1007 12:25:29.542760 6817 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542778 6817 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542791 6817 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1007 12:25:29.542806 6817 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nF1007 12:25:29.542812 6817 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.300212 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.317570 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.331624 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.336192 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.336245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.336259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.336281 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.336295 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.351918 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.366915 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.385786 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.398496 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.413534 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.430064 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.439399 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.439461 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.439479 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.439507 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.439523 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.447735 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.463995 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.478879 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.492866 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.505991 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:31Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.542076 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.542146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.542191 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.542223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.542243 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.640141 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:31 crc kubenswrapper[4702]: E1007 12:25:31.640441 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.644490 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.644529 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.644538 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.644553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.644563 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.747788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.747842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.747855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.747874 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.747911 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.850762 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.851463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.851502 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.851525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.851540 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.954530 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.954579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.954592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.954618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:31 crc kubenswrapper[4702]: I1007 12:25:31.954632 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:31Z","lastTransitionTime":"2025-10-07T12:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.058691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.058763 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.058789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.058822 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.058846 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.162859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.162964 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.162987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.163016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.163037 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.266176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.266228 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.266244 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.266266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.266280 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.369050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.369108 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.369124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.369177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.369195 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.473500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.473598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.473629 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.473668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.473690 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.577036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.577095 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.577109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.577129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.577143 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.640257 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.640314 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.640455 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.640551 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.640653 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.640775 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.680719 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.680763 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.680777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.680798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.680814 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.784685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.784767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.784794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.784825 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.784850 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.835601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.835675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.835687 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.835710 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.835723 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.854844 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.859803 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.859890 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.859942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.859973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.859997 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.878565 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.884607 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.884654 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.884666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.884723 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.884736 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.902482 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.907351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.907402 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.907415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.907435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.907446 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.926051 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.931304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.931338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.931350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.931391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.931404 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.949937 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:32Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:32 crc kubenswrapper[4702]: E1007 12:25:32.950141 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.952404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.952451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.952462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.952482 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:32 crc kubenswrapper[4702]: I1007 12:25:32.952493 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:32Z","lastTransitionTime":"2025-10-07T12:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.056303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.056354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.056365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.056388 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.056399 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.159502 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.159586 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.159611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.159643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.159672 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.262534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.262598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.262616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.262643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.262665 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.365975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.366039 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.366050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.366070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.366080 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.469554 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.469632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.469651 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.469684 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.469707 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.573352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.573426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.573443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.573469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.573485 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.640437 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:33 crc kubenswrapper[4702]: E1007 12:25:33.640653 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.677236 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.677300 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.677312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.677333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.677350 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.780962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.781009 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.781018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.781040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.781052 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.884316 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.884366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.884376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.884397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.884410 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.987450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.987514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.987532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.987559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:33 crc kubenswrapper[4702]: I1007 12:25:33.987581 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:33Z","lastTransitionTime":"2025-10-07T12:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.090348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.090419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.090432 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.090455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.090465 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.193503 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.193584 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.193603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.193630 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.193649 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.296767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.296821 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.296832 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.296853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.296867 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.400294 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.400359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.400378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.400405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.400426 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.503410 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.503486 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.503510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.503542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.503565 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.606948 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.607018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.607035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.607073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.607091 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.641213 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.641296 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.641442 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:34 crc kubenswrapper[4702]: E1007 12:25:34.641448 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:34 crc kubenswrapper[4702]: E1007 12:25:34.641876 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:34 crc kubenswrapper[4702]: E1007 12:25:34.642047 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.662651 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.685121 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.709776 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.710940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.711000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.711017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.711050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.711071 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.736721 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.756952 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.775134 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.798767 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.815168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.815240 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.815266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.815295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.815319 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.826225 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.850759 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.870479 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.885482 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.895020 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.906679 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.917401 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.919716 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.919769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.919783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.919807 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.919823 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:34Z","lastTransitionTime":"2025-10-07T12:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.930548 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.945421 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.967427 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:29Z\\\",\\\"message\\\":\\\"anager/kube-controller-manager-crc openshift-multus/network-metrics-daemon-znndd openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-multus/multus-additional-cni-plugins-t9vtm openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1007 12:25:29.542730 6817 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1007 12:25:29.542760 6817 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542778 6817 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542791 6817 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1007 12:25:29.542806 6817 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nF1007 12:25:29.542812 6817 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:34 crc kubenswrapper[4702]: I1007 12:25:34.979881 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:34Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.023083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.023139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.023156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.023181 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.023201 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.126715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.126791 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.126813 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.126844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.126864 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.229773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.229843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.229864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.229919 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.229941 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.333662 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.333772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.333785 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.333809 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.333825 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.436777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.436842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.436859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.436886 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.436940 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.541504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.541553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.541568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.541591 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.541607 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.640130 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:35 crc kubenswrapper[4702]: E1007 12:25:35.640316 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.644029 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.644084 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.644105 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.644135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.644157 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.747308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.747350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.747361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.747380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.747392 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.850638 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.851064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.851257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.851398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.851518 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.955163 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.956029 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.956061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.956081 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:35 crc kubenswrapper[4702]: I1007 12:25:35.956136 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:35Z","lastTransitionTime":"2025-10-07T12:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.059458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.059537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.059558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.059583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.059599 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.162611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.162679 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.162702 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.162730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.162750 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.266454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.266524 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.266545 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.266578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.266599 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.369809 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.369929 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.369956 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.369989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.370017 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.474435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.474531 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.474607 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.474639 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.474661 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.578656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.578724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.578743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.578771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.578792 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.640230 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.640337 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.640358 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:36 crc kubenswrapper[4702]: E1007 12:25:36.640565 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:36 crc kubenswrapper[4702]: E1007 12:25:36.640605 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:36 crc kubenswrapper[4702]: E1007 12:25:36.640733 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.681667 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.681701 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.681710 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.681727 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.681737 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.784641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.784678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.784689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.784709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.784721 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.887439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.887488 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.887497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.887515 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.887528 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.991253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.991333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.991354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.991381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:36 crc kubenswrapper[4702]: I1007 12:25:36.991399 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:36Z","lastTransitionTime":"2025-10-07T12:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.094853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.094923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.094934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.094952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.094963 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.198136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.198179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.198192 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.198213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.198225 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.302691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.303295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.303451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.303601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.303753 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.407499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.407552 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.407562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.407582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.407593 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.511052 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.511159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.511172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.511195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.511208 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.614799 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.614855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.614868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.614886 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.614917 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.640781 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:37 crc kubenswrapper[4702]: E1007 12:25:37.641025 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.718819 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.718874 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.718887 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.718932 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.718944 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.822771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.822834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.822847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.822871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.822886 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.926754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.926817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.926833 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.926864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:37 crc kubenswrapper[4702]: I1007 12:25:37.926882 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:37Z","lastTransitionTime":"2025-10-07T12:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.029767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.029817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.029827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.029843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.029852 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.132132 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.132166 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.132175 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.132194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.132203 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.236011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.236045 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.236054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.236071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.236081 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.339773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.339839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.339862 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.339941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.339963 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.443613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.443661 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.443674 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.443692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.443703 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.547066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.547104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.547115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.547136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.547146 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.640284 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:38 crc kubenswrapper[4702]: E1007 12:25:38.640538 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.640283 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.640283 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:38 crc kubenswrapper[4702]: E1007 12:25:38.640667 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:38 crc kubenswrapper[4702]: E1007 12:25:38.640917 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.648864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.648950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.648968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.648990 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.649007 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.753030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.753103 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.753123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.753150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.753170 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.856578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.856634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.856652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.856674 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.856688 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.961402 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.961475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.961494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.961525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:38 crc kubenswrapper[4702]: I1007 12:25:38.961545 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:38Z","lastTransitionTime":"2025-10-07T12:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.065072 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.065114 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.065123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.065140 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.065150 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.168987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.169061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.169079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.169135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.169153 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.272257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.272305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.272320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.272337 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.272351 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.375418 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.375481 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.375492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.375510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.375519 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.478080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.478157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.478180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.478210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.478230 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.581118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.581203 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.581217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.581241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.581255 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.640709 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:39 crc kubenswrapper[4702]: E1007 12:25:39.640888 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.684305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.684376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.684396 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.684422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.684443 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.787466 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.787549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.787574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.787608 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.787631 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.891335 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.891390 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.891399 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.891417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.891434 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.994500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.994565 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.994588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.994620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:39 crc kubenswrapper[4702]: I1007 12:25:39.994643 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:39Z","lastTransitionTime":"2025-10-07T12:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.097257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.097304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.097318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.097599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.097658 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.201582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.201657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.201675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.201705 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.201732 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.304689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.304827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.304862 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.304888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.304940 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.408698 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.408745 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.408756 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.408774 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.408785 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.512169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.512255 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.512275 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.512307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.512331 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.615220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.615274 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.615286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.615305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.615319 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.640151 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.640295 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:40 crc kubenswrapper[4702]: E1007 12:25:40.640379 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.640192 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:40 crc kubenswrapper[4702]: E1007 12:25:40.640566 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:40 crc kubenswrapper[4702]: E1007 12:25:40.640841 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.657266 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.718936 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.719022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.719059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.719083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.719100 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.821460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.821537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.821560 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.821590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.821614 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.925548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.925603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.925621 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.925645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:40 crc kubenswrapper[4702]: I1007 12:25:40.925664 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:40Z","lastTransitionTime":"2025-10-07T12:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.030534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.030599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.030621 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.030648 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.030669 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.133499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.133566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.133577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.133610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.133621 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.237559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.237641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.237656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.237678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.237691 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.340445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.340545 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.340555 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.340577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.340589 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.443460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.443497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.443506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.443522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.443532 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.546526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.546637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.546658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.546694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.546715 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.640843 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:41 crc kubenswrapper[4702]: E1007 12:25:41.641071 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.649964 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.650011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.650027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.650045 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.650058 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.753049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.753112 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.753130 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.753149 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.753163 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.856095 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.856180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.856205 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.856241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.856260 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.958454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.958497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.958510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.958528 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:41 crc kubenswrapper[4702]: I1007 12:25:41.958540 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:41Z","lastTransitionTime":"2025-10-07T12:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.061288 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.061332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.061347 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.061386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.061401 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.164601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.164652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.164864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.164888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.164934 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.268125 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.268180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.268190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.268209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.268223 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.371493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.371565 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.371583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.371609 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.371628 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.474617 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.474669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.474679 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.474694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.474704 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.577598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.577682 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.577696 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.577749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.577766 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.640724 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.640751 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.640724 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:42 crc kubenswrapper[4702]: E1007 12:25:42.640886 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:42 crc kubenswrapper[4702]: E1007 12:25:42.640970 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:42 crc kubenswrapper[4702]: E1007 12:25:42.641044 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.680787 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.680830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.680846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.680868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.680885 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.783981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.784054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.784073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.784102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.784121 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.888020 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.888118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.888150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.888185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.888209 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.992994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.993047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.993056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.993079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:42 crc kubenswrapper[4702]: I1007 12:25:42.993091 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:42Z","lastTransitionTime":"2025-10-07T12:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.097204 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.097275 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.097299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.097330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.097351 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.193053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.193156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.193185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.193220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.193243 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.217431 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.223253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.223309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.223322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.223346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.223362 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.239002 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.250447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.250529 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.250571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.250597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.250612 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.271329 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.277129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.277228 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.277258 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.277295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.277322 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.296371 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.301232 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.301311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.301340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.301376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.301403 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.318817 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:43Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.319025 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.320916 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.320951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.320964 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.320984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.321000 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.424024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.424080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.424103 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.424147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.424161 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.527254 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.527326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.527338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.527364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.527378 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.629531 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.629577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.629587 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.629603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.629618 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.640166 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.640400 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.641418 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:25:43 crc kubenswrapper[4702]: E1007 12:25:43.641740 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.732656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.732768 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.732782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.732801 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.732814 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.836303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.836772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.836885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.837088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.837201 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.940770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.940920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.940941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.940973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:43 crc kubenswrapper[4702]: I1007 12:25:43.940993 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:43Z","lastTransitionTime":"2025-10-07T12:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.044538 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.044940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.045019 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.045085 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.045155 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.148262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.148854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.149139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.149336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.149497 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.254523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.254595 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.254614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.254643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.254662 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.358448 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.358986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.359199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.359410 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.359607 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.463457 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.463524 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.463536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.463558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.463577 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.567565 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.567617 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.567632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.567656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.567669 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.640286 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.640458 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:44 crc kubenswrapper[4702]: E1007 12:25:44.640555 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:44 crc kubenswrapper[4702]: E1007 12:25:44.640603 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.640787 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:44 crc kubenswrapper[4702]: E1007 12:25:44.640924 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.657793 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13e19b0f-0b95-4a8c-a879-e60d42ea7f75\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25d1b57e56fd1779e665217ba711e25a7bb3f08c62b580511149a6bd60531ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24554634f1020aee5f8d32b27f261dd8edf9c691156b8f3d6f32cb850396849d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54022cfb6eba7182a7ace2b3a36a4931b0af004772d5dca69da5baf152cf040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3229afac41c60d0eedf1d935805d299a56330c0dd2390a65e27f56cde227be0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.670776 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.670847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.670866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.670934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.670961 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.685378 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f045d6-402b-4674-8ef6-2c8b74aa83ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52e1ae7c219f810a11ad287256dfa460a2a85ee4873ef407410d1e9c7e7e6b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef76a200b923258abceee8b6c914b1bbfff49fbe47f9ba9083e8e6201b908633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f0cce1ebdc322c8367466f6396631e74f4991948ec81f4051512d398026c40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80396ff718cf571bd3f3ac0a4439ec2b73801e0763a79130bb5eda3c43a03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16bbf92ad960448a44a3547670cf7f14105e8fae8f1a359c53621c331cd304f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b86cf396b84e1a8a67c39d95f5c06c283ef3c873bbd8384e8cddef2b770fdb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b86cf396b84e1a8a67c39d95f5c06c283ef3c873bbd8384e8cddef2b770fdb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f272bee882c6210f9fe9f5af37df7965e8bd3e773e29027ca0ee53496e110c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f272bee882c6210f9fe9f5af37df7965e8bd3e773e29027ca0ee53496e110c9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://011b7ff0a70a867cf2bbe100dbf1ebafe40b410576129e09dfa958014c0b3d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011b7ff0a70a867cf2bbe100dbf1ebafe40b410576129e09dfa958014c0b3d86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.700849 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5f3003-5108-4636-adb3-a691d977cace\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd5736a92bc038d8c0d9871c84f88730817c76275a4f0f73f1edf2712bf1c1b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a363cd4c8a004817654b9dade3fe311a8cbe61cb8937ebeaff9dbbcc519c55cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5c2e1e7750bc0790dfe66f506aadfa298c0a8e1009a05f163630a646ee64c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ed73eed98e252c6399abf92aad372546eff16a17bbf22d16bcc052374f5a9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6289ac9e015207691c44b7210e0000816a504cd2dc00ddb5bfde4f3bbf94846a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"ES_128_CBC_SHA256' detected.\\\\nW1007 12:24:24.402520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 12:24:24.402528 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 12:24:24.402534 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 12:24:24.402538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1007 12:24:24.402515 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1007 12:24:24.406313 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406355 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1007 12:24:24.406373 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1007 12:24:24.406400 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF1007 12:24:24.407100 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 12:24:24.408972 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1007 12:24:24.409004 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1007 12:24:24.406396 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1007 12:24:24.409396 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1007 12:24:24.409618 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1125301850/tls.crt::/tmp/serving-cert-1125301850/tls.key\\\\\\\"\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9347901bf7674e80879334c16a746c37f229af38027b5b8b5394c2ce02fdf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa1b13a13e0769bea92bcce6e837611473fbbe74ba696fbff1a77c6a5d573f12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.714872 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b681370383a37bd0f47713ee733d14fa0f327251f78bc769129a9ade16490dfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5362a96b2535f4c47fbb3b018adbea372dc83c1c22a60d6fbec9fba0eee689f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.733107 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf8728e-da3d-4d2a-9b9b-6094b460f50b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfc85cb927cbc0e021967ad4f0d36ee5b7334d276d11d65125d3e8e47c6f5421\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07b8fa7cf1f5c12f37ab224e4640fb52b4d82a8f0f867afe5b1b3312f992b90\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://695145edf3ffb661376e54c1881bb8846b0ed0aa7679e6d8d25cfb312611782d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d94ad663d8fa5ee93921fbb8645d2c2f2e2fc88071d2db6cc015fc1253e8ccda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b7bf48eef5d318d5f7991585260d727b2e38de3a47f466ad86d6fcff8e34e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37d9080ffff23dfb12650e3ba3fe5b1a4ec37489e9a4160753faf147142b5dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9428cdfcb827ed31cee75b45a6809d240223e5758148cf279a5e0e59c1c5e8af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mqgzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9vtm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.749470 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1ca41d2-75d4-4871-abe5-46a3e172fadd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8707bd812e59ad239a58fa1de731bbc21a5d3d26fb36656992bb8a847559e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w4btn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.769213 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-znndd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fb0cea3-d8a9-4738-abf3-47eea11213d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vczpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-znndd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.773769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.773839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.773874 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.773921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.773935 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.787018 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce54277cbcbf29b028e7c571b1e4c1321e8e0f274259e272793e383e1eea0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.803635 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.817501 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56edcf622272a048712586fd1a5f857626dcf8ac9a4728d73e515aad524e8ce3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.829307 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jrz4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d66a83d-539e-4a9b-8299-bf79fcd54629\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://324d7cbc29fa4acc89949e99fbf6ebe49956897a9b420a8cfabee8678dd3b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4htvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jrz4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.843167 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad6b96be-e641-4d3e-a8d8-a750d029fdc5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d714198c855a68ce0ecff08bcca877339be4e7db47209db69ee42031ddc73c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a26c4af17436fca094f2356685dad32bf731fea59021befcf0b17c357bf046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xxh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f4cq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.861837 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"867ec462-1be2-4d8e-936c-a1e40498494c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82a1660f89705f018df8a0b843cc2c7549d6bff213f21710e5da1f9b11e65258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b2e17042a303469479f73da5237bfb334759151314d2522ff4abe8dd670585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e906749ea81a6a5a1a9e01d6dbdfc8173a340c5bd4867d5174cad9245624a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ac57777e6c2c1ebbd587b0a3c8111e491d078b767f62b7dbb7dbfd6bdd822b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.872108 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bad91b4-b4c7-4d9a-bcf1-09b9a2935db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2070c24aa8d7f4092f0af7c2bea658c7ce5c8df374afc08227e35575965116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e500ee2f87ce3258e527f0ccd3f79ce39a2a447fa4bd07a4b9505e91be5b402\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.876095 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.876134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.876145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.876164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.876176 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.888109 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.905527 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.917799 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-djf9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f302f84b-82f7-4c73-a179-e7c4e20d05c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a43d841051644f12cc7e3e2ad13d509f3407e0c74fe762cf024e27b4c194dde3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d7h6p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-djf9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.935272 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9s7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"943865b8-ec83-4c7e-9e71-b02eaea71c1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:18Z\\\",\\\"message\\\":\\\"2025-10-07T12:24:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f\\\\n2025-10-07T12:24:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1465d67c-3df9-4fe6-a2b6-82b765c6544f to /host/opt/cni/bin/\\\\n2025-10-07T12:24:33Z [verbose] multus-daemon started\\\\n2025-10-07T12:24:33Z [verbose] Readiness Indicator file check\\\\n2025-10-07T12:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m5xqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9s7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.955590 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b62ec2-6971-4e45-bff9-5b65b657b6b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T12:25:29Z\\\",\\\"message\\\":\\\"anager/kube-controller-manager-crc openshift-multus/network-metrics-daemon-znndd openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-multus/multus-additional-cni-plugins-t9vtm openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1007 12:25:29.542730 6817 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1007 12:25:29.542760 6817 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542778 6817 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1007 12:25:29.542791 6817 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1007 12:25:29.542806 6817 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nF1007 12:25:29.542812 6817 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T12:25:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T12:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T12:24:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T12:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d2c9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T12:24:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pcg5c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:44Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.978913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.978954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.978966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.978985 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:44 crc kubenswrapper[4702]: I1007 12:25:44.979001 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:44Z","lastTransitionTime":"2025-10-07T12:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.082350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.082451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.082519 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.082544 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.082555 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.187179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.187258 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.187276 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.187308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.187328 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.289677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.289736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.289753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.289778 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.289794 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.392312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.392367 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.392378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.392396 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.392409 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.494999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.495050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.495059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.495078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.495092 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.597982 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.598060 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.598078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.598104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.598122 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.640748 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:45 crc kubenswrapper[4702]: E1007 12:25:45.641583 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.700859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.700952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.700969 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.700994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.701008 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.804250 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.804300 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.804312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.804331 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.804343 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.907197 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.907245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.907262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.907283 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:45 crc kubenswrapper[4702]: I1007 12:25:45.907296 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:45Z","lastTransitionTime":"2025-10-07T12:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.010132 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.010458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.010526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.010611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.010670 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.114021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.114382 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.114498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.114605 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.114711 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.218207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.218294 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.218318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.218351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.218373 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.321785 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.321845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.321855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.321875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.321889 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.425086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.425139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.425156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.425182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.425207 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.528281 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.528345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.528360 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.528381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.528397 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.631846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.631983 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.632003 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.632031 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.632053 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.640477 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.640634 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:46 crc kubenswrapper[4702]: E1007 12:25:46.640658 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:46 crc kubenswrapper[4702]: E1007 12:25:46.641110 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.641289 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:46 crc kubenswrapper[4702]: E1007 12:25:46.641398 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.734709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.734776 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.734793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.734821 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.734838 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.837498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.837542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.837551 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.837595 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.837612 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.941444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.941512 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.941527 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.941549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:46 crc kubenswrapper[4702]: I1007 12:25:46.941564 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:46Z","lastTransitionTime":"2025-10-07T12:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.044883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.044963 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.044979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.045000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.045012 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.148344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.148383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.148395 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.148414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.148425 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.251675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.251747 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.251761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.251782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.251795 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.354619 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.354667 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.354679 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.354699 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.354709 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.457127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.457168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.457176 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.457193 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.457204 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.559758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.559806 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.559824 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.559843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.559854 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.640523 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:47 crc kubenswrapper[4702]: E1007 12:25:47.640687 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.662199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.662241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.662255 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.662279 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.662292 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.764680 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.764724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.764736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.764754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.764766 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.867353 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.867393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.867406 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.867447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.867461 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.970369 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.970408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.970417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.970437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.970446 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:47Z","lastTransitionTime":"2025-10-07T12:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:47 crc kubenswrapper[4702]: I1007 12:25:47.981280 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:47 crc kubenswrapper[4702]: E1007 12:25:47.981480 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:25:47 crc kubenswrapper[4702]: E1007 12:25:47.981570 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs podName:8fb0cea3-d8a9-4738-abf3-47eea11213d7 nodeName:}" failed. No retries permitted until 2025-10-07 12:26:51.981550574 +0000 UTC m=+168.157286690 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs") pod "network-metrics-daemon-znndd" (UID: "8fb0cea3-d8a9-4738-abf3-47eea11213d7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.073339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.073366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.073380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.073397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.073409 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.176070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.176121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.176136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.176155 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.176165 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.278066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.278126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.278137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.278152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.278161 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.380783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.380824 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.380832 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.380851 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.380861 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.484338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.484397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.484414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.484441 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.484455 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.587664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.587709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.587722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.587744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.587757 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.640630 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:48 crc kubenswrapper[4702]: E1007 12:25:48.640810 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.640888 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:48 crc kubenswrapper[4702]: E1007 12:25:48.640986 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.641166 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:48 crc kubenswrapper[4702]: E1007 12:25:48.641283 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.690979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.691043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.691058 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.691083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.691098 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.794356 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.794412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.794428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.794454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.794469 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.896928 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.896967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.896979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.896994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.897003 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:48 crc kubenswrapper[4702]: I1007 12:25:48.999658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:48.999705 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:48.999753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:48.999776 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:48.999787 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:48Z","lastTransitionTime":"2025-10-07T12:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.103429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.103480 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.103489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.103508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.103517 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.207194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.207249 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.207262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.207284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.207292 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.310141 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.310184 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.310196 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.310213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.310224 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.413210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.413263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.413274 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.413296 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.413306 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.516298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.516369 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.516381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.516403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.516414 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.618571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.618632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.618646 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.618666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.618678 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.640314 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:49 crc kubenswrapper[4702]: E1007 12:25:49.640493 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.721827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.721866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.721877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.721893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.721924 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.824262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.824312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.824324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.824357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.824368 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.926355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.926396 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.926405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.926422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:49 crc kubenswrapper[4702]: I1007 12:25:49.926432 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:49Z","lastTransitionTime":"2025-10-07T12:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.029814 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.029881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.029911 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.029939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.029955 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.133147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.133198 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.133247 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.133275 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.133288 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.236198 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.236247 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.236257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.236276 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.236289 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.339278 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.339312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.339322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.339340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.339353 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.442243 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.442307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.442323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.442348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.442370 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.545787 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.545841 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.545854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.545873 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.545888 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.640663 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.640745 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.640761 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:50 crc kubenswrapper[4702]: E1007 12:25:50.640875 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:50 crc kubenswrapper[4702]: E1007 12:25:50.641139 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:50 crc kubenswrapper[4702]: E1007 12:25:50.641253 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.649124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.649198 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.649217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.649246 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.649266 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.752410 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.752463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.752481 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.752505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.752521 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.855836 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.855885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.855914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.855930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.855941 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.959788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.959884 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.959950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.959986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:50 crc kubenswrapper[4702]: I1007 12:25:50.960014 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:50Z","lastTransitionTime":"2025-10-07T12:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.063736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.063790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.063798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.063817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.063827 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.166658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.166724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.166736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.166761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.166777 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.269318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.269360 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.269371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.269386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.269401 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.372656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.372716 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.372727 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.372748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.372761 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.475704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.475753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.475765 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.475783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.475793 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.579322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.579365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.579374 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.579392 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.579401 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.640396 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:51 crc kubenswrapper[4702]: E1007 12:25:51.640648 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.683430 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.683505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.683523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.683554 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.683576 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.786403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.786461 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.786479 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.786508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.786527 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.890100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.890136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.890146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.890167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.890181 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.993486 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.993527 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.993537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.993556 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:51 crc kubenswrapper[4702]: I1007 12:25:51.993569 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:51Z","lastTransitionTime":"2025-10-07T12:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.098052 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.098121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.098135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.098161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.098177 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.201368 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.201429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.201445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.201469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.201485 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.303993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.304064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.304077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.304097 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.304107 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.406397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.406467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.406480 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.406498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.406508 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.509505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.509567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.509579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.509604 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.509619 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.611819 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.611854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.611868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.611891 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.611927 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.640065 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:52 crc kubenswrapper[4702]: E1007 12:25:52.640216 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.640375 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:52 crc kubenswrapper[4702]: E1007 12:25:52.640563 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.640708 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:52 crc kubenswrapper[4702]: E1007 12:25:52.640884 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.714954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.715033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.715050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.715073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.715090 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.817937 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.818011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.818029 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.818056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.818077 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.920408 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.920476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.920488 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.920510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:52 crc kubenswrapper[4702]: I1007 12:25:52.920547 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:52Z","lastTransitionTime":"2025-10-07T12:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.022772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.022804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.022813 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.022830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.022839 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.125800 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.125833 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.125843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.125860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.125870 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.227826 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.227866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.227878 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.227939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.227964 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.330065 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.330107 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.330119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.330137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.330150 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.432749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.432797 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.432812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.432834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.432848 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.514713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.514769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.514781 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.514800 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.514814 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.530073 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:53Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.535870 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.535943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.535959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.535979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.535992 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.550992 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:53Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.555855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.555977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.556004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.556037 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.556062 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.574003 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:53Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.578811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.578858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.578871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.578913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.578926 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.592549 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:53Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.597637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.597689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.597700 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.597733 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.597746 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.610425 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T12:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"765d68a7-0b5e-4d6a-980f-84924188d680\\\",\\\"systemUUID\\\":\\\"bbdbab7f-ca5e-4664-b776-056ee45e0f4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T12:25:53Z is after 2025-08-24T17:21:41Z" Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.610635 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.613476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.613537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.613561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.613595 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.613619 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.640344 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:53 crc kubenswrapper[4702]: E1007 12:25:53.640794 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.716721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.716813 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.716839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.716869 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.716891 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.819510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.819592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.819611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.819643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.819712 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.922087 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.922125 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.922134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.922151 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:53 crc kubenswrapper[4702]: I1007 12:25:53.922161 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:53Z","lastTransitionTime":"2025-10-07T12:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.024731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.024768 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.024777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.024794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.024804 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.131528 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.131652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.131679 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.131717 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.131755 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.235063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.235130 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.235148 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.235173 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.235189 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.338146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.338207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.338218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.338237 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.338248 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.440955 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.440997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.441013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.441037 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.441055 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.544118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.544167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.544177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.544199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.544209 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.640651 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.640715 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.640651 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:54 crc kubenswrapper[4702]: E1007 12:25:54.640847 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:54 crc kubenswrapper[4702]: E1007 12:25:54.641113 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:54 crc kubenswrapper[4702]: E1007 12:25:54.641162 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.647056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.647109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.647123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.647143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.647156 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.699681 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-djf9f" podStartSLOduration=85.699648348 podStartE2EDuration="1m25.699648348s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.699648298 +0000 UTC m=+110.875384414" watchObservedRunningTime="2025-10-07 12:25:54.699648348 +0000 UTC m=+110.875384474" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.699876 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2jrz4" podStartSLOduration=85.699870385 podStartE2EDuration="1m25.699870385s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.685515436 +0000 UTC m=+110.861251552" watchObservedRunningTime="2025-10-07 12:25:54.699870385 +0000 UTC m=+110.875606511" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.740529 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h9s7h" podStartSLOduration=85.740500732 podStartE2EDuration="1m25.740500732s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.714877352 +0000 UTC m=+110.890613468" watchObservedRunningTime="2025-10-07 12:25:54.740500732 +0000 UTC m=+110.916236858" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.749479 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.749540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.749550 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.749566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.749576 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.783259 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f4cq9" podStartSLOduration=85.783212069 podStartE2EDuration="1m25.783212069s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.75762067 +0000 UTC m=+110.933356806" watchObservedRunningTime="2025-10-07 12:25:54.783212069 +0000 UTC m=+110.958948195" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.783953 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=90.78394656 podStartE2EDuration="1m30.78394656s" podCreationTimestamp="2025-10-07 12:24:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.782943901 +0000 UTC m=+110.958680017" watchObservedRunningTime="2025-10-07 12:25:54.78394656 +0000 UTC m=+110.959682686" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.815914 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=42.815866559 podStartE2EDuration="42.815866559s" podCreationTimestamp="2025-10-07 12:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.80009423 +0000 UTC m=+110.975830346" watchObservedRunningTime="2025-10-07 12:25:54.815866559 +0000 UTC m=+110.991602665" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.852857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.852937 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.852949 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.852976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.852990 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.865617 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-t9vtm" podStartSLOduration=85.865565115 podStartE2EDuration="1m25.865565115s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.86364333 +0000 UTC m=+111.039379446" watchObservedRunningTime="2025-10-07 12:25:54.865565115 +0000 UTC m=+111.041301251" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.881062 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podStartSLOduration=85.881035485 podStartE2EDuration="1m25.881035485s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.879545873 +0000 UTC m=+111.055281999" watchObservedRunningTime="2025-10-07 12:25:54.881035485 +0000 UTC m=+111.056771621" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.906287 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.906265044 podStartE2EDuration="56.906265044s" podCreationTimestamp="2025-10-07 12:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.905610495 +0000 UTC m=+111.081346621" watchObservedRunningTime="2025-10-07 12:25:54.906265044 +0000 UTC m=+111.082001160" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956233 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956287 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956300 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956337 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:54Z","lastTransitionTime":"2025-10-07T12:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956489 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=14.956467794 podStartE2EDuration="14.956467794s" podCreationTimestamp="2025-10-07 12:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.937337679 +0000 UTC m=+111.113073795" watchObservedRunningTime="2025-10-07 12:25:54.956467794 +0000 UTC m=+111.132203910" Oct 07 12:25:54 crc kubenswrapper[4702]: I1007 12:25:54.956794 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.956788983 podStartE2EDuration="1m30.956788983s" podCreationTimestamp="2025-10-07 12:24:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:25:54.956093914 +0000 UTC m=+111.131830040" watchObservedRunningTime="2025-10-07 12:25:54.956788983 +0000 UTC m=+111.132525099" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.059959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.060023 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.060036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.060062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.060078 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.162790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.162855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.162868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.162891 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.162925 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.264973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.265011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.265022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.265042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.265053 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.367574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.367618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.367627 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.367647 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.368023 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.470966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.471032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.471044 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.471063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.471076 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.574175 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.574216 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.574230 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.574251 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.574264 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.640278 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:55 crc kubenswrapper[4702]: E1007 12:25:55.640503 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.677104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.677154 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.677166 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.677188 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.677202 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.780241 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.780287 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.780298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.780319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.780330 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.883561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.883639 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.883662 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.883689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.883712 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.989088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.989173 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.989192 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.989415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:55 crc kubenswrapper[4702]: I1007 12:25:55.989563 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:55Z","lastTransitionTime":"2025-10-07T12:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.094928 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.095180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.095240 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.095271 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.095293 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.197970 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.198026 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.198036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.198055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.198068 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.300169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.300215 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.300229 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.300247 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.300257 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.403177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.403220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.403229 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.403246 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.403258 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.505777 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.505852 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.505864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.505887 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.505929 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.609564 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.609651 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.609680 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.609715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.609740 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.640050 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.640058 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:56 crc kubenswrapper[4702]: E1007 12:25:56.640341 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.640058 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:56 crc kubenswrapper[4702]: E1007 12:25:56.640529 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:56 crc kubenswrapper[4702]: E1007 12:25:56.640778 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.713009 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.713079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.713122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.713161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.713188 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.816016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.816071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.816086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.816105 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.816117 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.919585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.919634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.919646 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.919665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:56 crc kubenswrapper[4702]: I1007 12:25:56.919678 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:56Z","lastTransitionTime":"2025-10-07T12:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.023514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.023592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.023666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.023743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.023762 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.127351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.127423 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.127442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.127472 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.127493 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.230734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.230794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.230811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.230837 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.230855 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.334914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.334977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.334987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.335006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.335018 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.438085 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.438144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.438158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.438178 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.438230 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.541885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.541951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.541960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.541981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.541991 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.640229 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:57 crc kubenswrapper[4702]: E1007 12:25:57.640425 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.641332 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:25:57 crc kubenswrapper[4702]: E1007 12:25:57.641553 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pcg5c_openshift-ovn-kubernetes(68b62ec2-6971-4e45-bff9-5b65b657b6b6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.644934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.644968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.644983 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.645002 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.645017 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.747995 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.748074 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.748094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.748121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.748143 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.850888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.850995 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.851008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.851030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.851045 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.953936 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.953999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.954013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.954032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:57 crc kubenswrapper[4702]: I1007 12:25:57.954044 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:57Z","lastTransitionTime":"2025-10-07T12:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.056846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.056930 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.056940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.056958 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.056967 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.159329 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.159393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.159403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.159418 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.159428 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.261634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.261692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.261711 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.261732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.261745 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.364563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.364659 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.364686 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.364721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.364745 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.467730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.467798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.467820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.467844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.467860 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.571494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.571550 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.571566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.571592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.571605 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.640448 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.640456 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.640512 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:25:58 crc kubenswrapper[4702]: E1007 12:25:58.641106 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:25:58 crc kubenswrapper[4702]: E1007 12:25:58.641201 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:25:58 crc kubenswrapper[4702]: E1007 12:25:58.641391 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.674615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.674694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.674715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.674744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.674767 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.778501 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.778577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.778588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.778609 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.778619 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.881890 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.881954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.881967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.881987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.881999 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.984963 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.985031 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.985056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.985089 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:58 crc kubenswrapper[4702]: I1007 12:25:58.985116 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:58Z","lastTransitionTime":"2025-10-07T12:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.088390 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.088462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.088475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.088497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.088512 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.191522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.191573 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.191593 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.191614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.191627 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.294666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.294711 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.294721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.294740 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.294751 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.398021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.398066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.398078 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.398100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.398114 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.501242 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.501315 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.501327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.501345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.501375 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.604250 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.604309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.604321 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.604342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.604355 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.640275 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:25:59 crc kubenswrapper[4702]: E1007 12:25:59.640519 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.708168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.708252 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.708262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.708281 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.708291 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.811167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.811215 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.811224 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.811242 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.811252 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.914361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.914400 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.914412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.914433 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:25:59 crc kubenswrapper[4702]: I1007 12:25:59.914446 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:25:59Z","lastTransitionTime":"2025-10-07T12:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.017412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.017472 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.017491 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.017517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.017535 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.121131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.121212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.121233 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.121264 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.121305 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.224668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.224739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.224759 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.224785 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.224809 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.328958 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.329011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.329023 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.329042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.329053 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.434067 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.434571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.434723 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.434869 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.435090 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.538544 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.538620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.538644 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.538674 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.538699 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.641271 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.641330 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.641266 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:00 crc kubenswrapper[4702]: E1007 12:26:00.641549 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:00 crc kubenswrapper[4702]: E1007 12:26:00.641690 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:00 crc kubenswrapper[4702]: E1007 12:26:00.641816 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.642633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.642892 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.643186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.643439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.643688 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.747875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.748002 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.748036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.748071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.748098 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.852342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.852439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.852462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.852495 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.852510 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.955701 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.956005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.956138 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.956231 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:00 crc kubenswrapper[4702]: I1007 12:26:00.956341 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:00Z","lastTransitionTime":"2025-10-07T12:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.060110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.060175 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.060192 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.060217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.060237 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.163468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.163554 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.163571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.163596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.163611 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.267102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.267179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.267200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.267231 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.267253 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.370712 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.370802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.370826 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.370856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.370881 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.473865 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.474005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.474024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.474056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.474078 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.578059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.578592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.578719 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.578872 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.579061 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.639886 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:01 crc kubenswrapper[4702]: E1007 12:26:01.640472 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.682756 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.683189 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.683393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.683541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.683671 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.786424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.786486 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.786498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.786518 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.786528 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.889460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.889525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.889546 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.889579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.889606 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.992628 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.992710 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.992724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.992754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:01 crc kubenswrapper[4702]: I1007 12:26:01.992768 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:01Z","lastTransitionTime":"2025-10-07T12:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.096245 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.096301 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.096314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.096336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.096357 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.200068 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.200703 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.200724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.200752 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.200778 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.303965 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.304034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.304044 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.304063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.304073 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.407378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.407417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.407429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.407455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.407467 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.510321 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.510373 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.510383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.510402 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.510413 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.613647 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.613697 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.613709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.613729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.613749 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.640047 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.640126 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:02 crc kubenswrapper[4702]: E1007 12:26:02.640200 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:02 crc kubenswrapper[4702]: E1007 12:26:02.640325 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.640414 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:02 crc kubenswrapper[4702]: E1007 12:26:02.640506 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.716721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.716778 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.716796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.716820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.716839 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.819378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.819431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.819442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.819464 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.819480 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.923504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.923601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.923626 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.923692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:02 crc kubenswrapper[4702]: I1007 12:26:02.923714 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:02Z","lastTransitionTime":"2025-10-07T12:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.028119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.028195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.028219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.028253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.028282 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.132652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.132751 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.132771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.132831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.132862 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.237649 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.237772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.237808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.237849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.237892 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.340977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.341550 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.341762 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.341961 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.342153 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.445703 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.446319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.446487 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.446641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.446787 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.550879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.550990 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.551016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.551044 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.551062 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.640520 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:03 crc kubenswrapper[4702]: E1007 12:26:03.640694 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.653196 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.653225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.653234 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.653251 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.653263 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.755855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.756349 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.756553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.756747 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.756975 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.860330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.860381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.860394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.860415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.860427 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.887294 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.887340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.887352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.887371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.887384 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T12:26:03Z","lastTransitionTime":"2025-10-07T12:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.953121 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr"] Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.953620 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.957539 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.957732 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.957780 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.958333 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.973813 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.973868 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.973935 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.973960 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:03 crc kubenswrapper[4702]: I1007 12:26:03.974018 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075321 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075380 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075446 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075496 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075718 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.075743 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.077045 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.085552 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.109971 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e38f688-a98e-4a19-a73b-6c4ef9b606cc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lvhbr\" (UID: \"2e38f688-a98e-4a19-a73b-6c4ef9b606cc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.275244 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.330184 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" event={"ID":"2e38f688-a98e-4a19-a73b-6c4ef9b606cc","Type":"ContainerStarted","Data":"c9588a94773577a280c1067200bcd069ea2ab276328ac3d765d8af8323344577"} Oct 07 12:26:04 crc kubenswrapper[4702]: E1007 12:26:04.579613 4702 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.640139 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.640630 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:04 crc kubenswrapper[4702]: I1007 12:26:04.640674 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:04 crc kubenswrapper[4702]: E1007 12:26:04.641258 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:04 crc kubenswrapper[4702]: E1007 12:26:04.641539 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:04 crc kubenswrapper[4702]: E1007 12:26:04.641689 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:04 crc kubenswrapper[4702]: E1007 12:26:04.819810 4702 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.337617 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/1.log" Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.338388 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/0.log" Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.338489 4702 generic.go:334] "Generic (PLEG): container finished" podID="943865b8-ec83-4c7e-9e71-b02eaea71c1d" containerID="c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b" exitCode=1 Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.338870 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerDied","Data":"c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b"} Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.338999 4702 scope.go:117] "RemoveContainer" containerID="8e5148b72d204977b35d456518c3aac076d99c8b48acfe1f3021f64c6860d4fe" Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.339981 4702 scope.go:117] "RemoveContainer" containerID="c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b" Oct 07 12:26:05 crc kubenswrapper[4702]: E1007 12:26:05.340277 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h9s7h_openshift-multus(943865b8-ec83-4c7e-9e71-b02eaea71c1d)\"" pod="openshift-multus/multus-h9s7h" podUID="943865b8-ec83-4c7e-9e71-b02eaea71c1d" Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.340279 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" event={"ID":"2e38f688-a98e-4a19-a73b-6c4ef9b606cc","Type":"ContainerStarted","Data":"7e3a452f654dc558c7fbbc3d345463f1aaf98b5543c834736117a3bda7e112b6"} Oct 07 12:26:05 crc kubenswrapper[4702]: I1007 12:26:05.639752 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:05 crc kubenswrapper[4702]: E1007 12:26:05.640225 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:06 crc kubenswrapper[4702]: I1007 12:26:06.346178 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/1.log" Oct 07 12:26:06 crc kubenswrapper[4702]: I1007 12:26:06.640534 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:06 crc kubenswrapper[4702]: I1007 12:26:06.640586 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:06 crc kubenswrapper[4702]: I1007 12:26:06.640553 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:06 crc kubenswrapper[4702]: E1007 12:26:06.640747 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:06 crc kubenswrapper[4702]: E1007 12:26:06.640857 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:06 crc kubenswrapper[4702]: E1007 12:26:06.640980 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:07 crc kubenswrapper[4702]: I1007 12:26:07.640341 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:07 crc kubenswrapper[4702]: E1007 12:26:07.640533 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:08 crc kubenswrapper[4702]: I1007 12:26:08.640396 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:08 crc kubenswrapper[4702]: I1007 12:26:08.640504 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:08 crc kubenswrapper[4702]: E1007 12:26:08.640580 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:08 crc kubenswrapper[4702]: I1007 12:26:08.640623 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:08 crc kubenswrapper[4702]: E1007 12:26:08.640807 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:08 crc kubenswrapper[4702]: E1007 12:26:08.640881 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:09 crc kubenswrapper[4702]: I1007 12:26:09.640204 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:09 crc kubenswrapper[4702]: E1007 12:26:09.640751 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:09 crc kubenswrapper[4702]: E1007 12:26:09.821693 4702 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 12:26:10 crc kubenswrapper[4702]: I1007 12:26:10.641106 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:10 crc kubenswrapper[4702]: I1007 12:26:10.641172 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:10 crc kubenswrapper[4702]: I1007 12:26:10.641231 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:10 crc kubenswrapper[4702]: E1007 12:26:10.641350 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:10 crc kubenswrapper[4702]: E1007 12:26:10.642034 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:10 crc kubenswrapper[4702]: E1007 12:26:10.642267 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:10 crc kubenswrapper[4702]: I1007 12:26:10.642788 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.366926 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/3.log" Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.370246 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerStarted","Data":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.370783 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.398834 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lvhbr" podStartSLOduration=102.398809178 podStartE2EDuration="1m42.398809178s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:05.384304936 +0000 UTC m=+121.560041052" watchObservedRunningTime="2025-10-07 12:26:11.398809178 +0000 UTC m=+127.574545294" Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.399361 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podStartSLOduration=102.399355324 podStartE2EDuration="1m42.399355324s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:11.397392448 +0000 UTC m=+127.573128564" watchObservedRunningTime="2025-10-07 12:26:11.399355324 +0000 UTC m=+127.575091440" Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.610344 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-znndd"] Oct 07 12:26:11 crc kubenswrapper[4702]: I1007 12:26:11.610557 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:11 crc kubenswrapper[4702]: E1007 12:26:11.610762 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:12 crc kubenswrapper[4702]: I1007 12:26:12.640652 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:12 crc kubenswrapper[4702]: I1007 12:26:12.640654 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:12 crc kubenswrapper[4702]: I1007 12:26:12.640890 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:12 crc kubenswrapper[4702]: E1007 12:26:12.641472 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:12 crc kubenswrapper[4702]: E1007 12:26:12.641600 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:12 crc kubenswrapper[4702]: E1007 12:26:12.641732 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:13 crc kubenswrapper[4702]: I1007 12:26:13.640767 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:13 crc kubenswrapper[4702]: E1007 12:26:13.641023 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:14 crc kubenswrapper[4702]: I1007 12:26:14.640378 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:14 crc kubenswrapper[4702]: I1007 12:26:14.640378 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:14 crc kubenswrapper[4702]: I1007 12:26:14.640565 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:14 crc kubenswrapper[4702]: E1007 12:26:14.642017 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:14 crc kubenswrapper[4702]: E1007 12:26:14.642132 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:14 crc kubenswrapper[4702]: E1007 12:26:14.642242 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:14 crc kubenswrapper[4702]: E1007 12:26:14.822588 4702 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 12:26:15 crc kubenswrapper[4702]: I1007 12:26:15.640291 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:15 crc kubenswrapper[4702]: E1007 12:26:15.640479 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:16 crc kubenswrapper[4702]: I1007 12:26:16.640860 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:16 crc kubenswrapper[4702]: I1007 12:26:16.640948 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:16 crc kubenswrapper[4702]: I1007 12:26:16.641167 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:16 crc kubenswrapper[4702]: E1007 12:26:16.641207 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:16 crc kubenswrapper[4702]: E1007 12:26:16.641433 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:16 crc kubenswrapper[4702]: E1007 12:26:16.641776 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:16 crc kubenswrapper[4702]: I1007 12:26:16.641990 4702 scope.go:117] "RemoveContainer" containerID="c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b" Oct 07 12:26:17 crc kubenswrapper[4702]: I1007 12:26:17.394321 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/1.log" Oct 07 12:26:17 crc kubenswrapper[4702]: I1007 12:26:17.394676 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerStarted","Data":"33efb0b3bfe61433e0cea825895177a26cf9b61dbcc4e7bc8f53636d1f58bbfd"} Oct 07 12:26:17 crc kubenswrapper[4702]: I1007 12:26:17.640262 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:17 crc kubenswrapper[4702]: E1007 12:26:17.640526 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:18 crc kubenswrapper[4702]: I1007 12:26:18.640096 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:18 crc kubenswrapper[4702]: I1007 12:26:18.640176 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:18 crc kubenswrapper[4702]: I1007 12:26:18.640337 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:18 crc kubenswrapper[4702]: E1007 12:26:18.640333 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 12:26:18 crc kubenswrapper[4702]: E1007 12:26:18.640512 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 12:26:18 crc kubenswrapper[4702]: E1007 12:26:18.640645 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 12:26:19 crc kubenswrapper[4702]: I1007 12:26:19.640986 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:19 crc kubenswrapper[4702]: E1007 12:26:19.641229 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-znndd" podUID="8fb0cea3-d8a9-4738-abf3-47eea11213d7" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.640044 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.640180 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.640266 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.644983 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.645343 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.645567 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 07 12:26:20 crc kubenswrapper[4702]: I1007 12:26:20.655224 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 07 12:26:21 crc kubenswrapper[4702]: I1007 12:26:21.640774 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:21 crc kubenswrapper[4702]: I1007 12:26:21.643877 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 07 12:26:21 crc kubenswrapper[4702]: I1007 12:26:21.644973 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.235596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.298201 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-b6wlp"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.298987 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.301419 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9lq9"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.302196 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.303239 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.303406 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.304768 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.305477 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.307637 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-q72r8"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.308223 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.309840 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.310521 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.310967 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.311180 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.312106 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.312370 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.315982 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.316592 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.316680 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.316832 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.316604 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.317263 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.317721 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.323742 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.323832 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.323745 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.324266 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.324631 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.324888 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.325059 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.324742 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.325496 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.325636 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.331398 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qtvlc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.332587 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.332829 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.339566 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.341294 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.342451 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.350157 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.357125 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.357570 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.358004 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.358515 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.358826 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.360683 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.361640 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.363568 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t56br"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.365628 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367737 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367651 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mz42v"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366060 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366220 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366331 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366349 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366545 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366608 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366646 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366719 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366747 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366802 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.366970 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367051 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367481 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367562 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367609 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.368580 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.367647 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.369518 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.374671 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lbqfj"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.374988 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375194 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375340 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9r9nf"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375490 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375699 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375709 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375792 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.375974 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.376278 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.376729 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.376839 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.376934 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.376743 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377067 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377025 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377032 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377041 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377335 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377486 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377604 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.377777 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.378137 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9k26f"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.378865 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.379929 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.380684 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t72fg"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.381084 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.381410 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.388488 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.390950 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vw7pc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.391246 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mbn6w"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.391945 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.392196 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.392517 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.392684 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.393218 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.393437 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.393584 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.393800 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.393970 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394083 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394237 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394355 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394471 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394581 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394685 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394792 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.394935 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395060 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395213 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395221 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395300 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395442 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395497 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.395570 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.396193 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.396630 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.397029 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.400846 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.404665 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.404983 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405012 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405185 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405252 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405380 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405633 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405882 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.405889 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.421095 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6gdcn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.425383 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.426241 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-n66bc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.430691 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-client-ca\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.431035 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-serving-cert\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.434607 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464255 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-oauth-serving-cert\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464308 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-oauth-config\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464333 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-config\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464354 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b28c51f3-e546-4bee-923f-e27b8c7c7567-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464370 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464392 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-trusted-ca-bundle\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464413 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-serving-cert\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464431 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-auth-proxy-config\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464445 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-config\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464461 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s258v\" (UniqueName: \"kubernetes.io/projected/b28c51f3-e546-4bee-923f-e27b8c7c7567-kube-api-access-s258v\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464491 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl78l\" (UniqueName: \"kubernetes.io/projected/e95bef62-5bd2-4a64-b1c5-df780f79ac51-kube-api-access-hl78l\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464510 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5abee0e6-194a-4457-86dc-6d27a745589b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464539 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464560 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-service-ca-bundle\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464575 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-config\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464593 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf2rd\" (UniqueName: \"kubernetes.io/projected/32ce1ea1-c409-4850-855b-fa8316d0d5e6-kube-api-access-rf2rd\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464609 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3ed3536-f707-47a5-b3df-3ec72bec773b-serving-cert\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464626 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-config\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464642 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-client-ca\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464667 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ce1ea1-c409-4850-855b-fa8316d0d5e6-serving-cert\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464686 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlsc7\" (UniqueName: \"kubernetes.io/projected/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-kube-api-access-qlsc7\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464709 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-machine-approver-tls\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464750 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqpk8\" (UniqueName: \"kubernetes.io/projected/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-kube-api-access-lqpk8\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464776 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5abee0e6-194a-4457-86dc-6d27a745589b-config\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464793 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-config\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464813 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5j4h\" (UniqueName: \"kubernetes.io/projected/5abee0e6-194a-4457-86dc-6d27a745589b-kube-api-access-p5j4h\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464830 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b28c51f3-e546-4bee-923f-e27b8c7c7567-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464848 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-service-ca\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464866 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5abee0e6-194a-4457-86dc-6d27a745589b-images\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.464886 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjpg\" (UniqueName: \"kubernetes.io/projected/a3ed3536-f707-47a5-b3df-3ec72bec773b-kube-api-access-vnjpg\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.465114 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.465871 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.466528 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.466872 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.468036 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.468710 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.469161 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.469486 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.488224 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.488619 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.488781 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.488877 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.489299 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.489774 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.489874 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.490126 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.492349 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.493329 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.494085 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.494169 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.494422 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.494873 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.495119 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.495143 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.499238 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.499420 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.500255 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-k4phl"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.500584 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.500684 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.500601 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcw2q"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.501318 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.501990 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.504955 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.505812 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.506280 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.506716 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.506802 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.506826 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.507032 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.507751 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.529738 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.542474 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.559668 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.560783 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.561608 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.562803 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9lq9"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.563804 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.564581 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565347 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qtvlc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565795 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-serving-cert\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565836 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78d2w\" (UniqueName: \"kubernetes.io/projected/a761b6de-3f12-4781-9a64-50bc79d6f482-kube-api-access-78d2w\") pod \"dns-operator-744455d44c-mbn6w\" (UID: \"a761b6de-3f12-4781-9a64-50bc79d6f482\") " pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565856 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtckm\" (UniqueName: \"kubernetes.io/projected/7758de03-0df5-45b4-8b7b-2a071740b2e0-kube-api-access-jtckm\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565878 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-config\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565916 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s258v\" (UniqueName: \"kubernetes.io/projected/b28c51f3-e546-4bee-923f-e27b8c7c7567-kube-api-access-s258v\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565937 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565954 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-auth-proxy-config\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565980 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl78l\" (UniqueName: \"kubernetes.io/projected/e95bef62-5bd2-4a64-b1c5-df780f79ac51-kube-api-access-hl78l\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.565997 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5abee0e6-194a-4457-86dc-6d27a745589b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566017 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k77m6\" (UniqueName: \"kubernetes.io/projected/7d163c22-5dce-4cae-b536-c86dabcf6a91-kube-api-access-k77m6\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566041 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566061 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566077 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-node-pullsecrets\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566093 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54366321-671a-46e6-a4ca-a834ba087059-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566109 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566124 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4cdf1f3-c551-4561-b731-a5413a13a724-metrics-tls\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566142 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/529c3d09-735a-4157-b35b-0adb74424387-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566164 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc49h\" (UniqueName: \"kubernetes.io/projected/529c3d09-735a-4157-b35b-0adb74424387-kube-api-access-lc49h\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566187 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-service-ca-bundle\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566208 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3ed3536-f707-47a5-b3df-3ec72bec773b-serving-cert\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566226 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-config\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566243 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-encryption-config\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566261 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7758de03-0df5-45b4-8b7b-2a071740b2e0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566276 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566294 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-config\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566313 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf2rd\" (UniqueName: \"kubernetes.io/projected/32ce1ea1-c409-4850-855b-fa8316d0d5e6-kube-api-access-rf2rd\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566328 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4cdf1f3-c551-4561-b731-a5413a13a724-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566343 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-config\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566359 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnrvv\" (UniqueName: \"kubernetes.io/projected/c207de79-b8c1-47f5-8e4b-5f320d760588-kube-api-access-pnrvv\") pod \"downloads-7954f5f757-9r9nf\" (UID: \"c207de79-b8c1-47f5-8e4b-5f320d760588\") " pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566377 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-config\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566393 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-client-ca\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566410 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566428 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-config\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566444 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-dir\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566461 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566478 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ce1ea1-c409-4850-855b-fa8316d0d5e6-serving-cert\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566494 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-etcd-client\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566510 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566527 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566553 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlsc7\" (UniqueName: \"kubernetes.io/projected/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-kube-api-access-qlsc7\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566570 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6gdcn\" (UID: \"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566590 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566606 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54366321-671a-46e6-a4ca-a834ba087059-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566620 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566638 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nq9x\" (UniqueName: \"kubernetes.io/projected/e4cdf1f3-c551-4561-b731-a5413a13a724-kube-api-access-6nq9x\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566655 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-machine-approver-tls\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566673 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566688 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-serving-cert\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566704 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x478z\" (UniqueName: \"kubernetes.io/projected/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-kube-api-access-x478z\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566720 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-client\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566747 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqpk8\" (UniqueName: \"kubernetes.io/projected/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-kube-api-access-lqpk8\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566763 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-audit-dir\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566778 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn25n\" (UniqueName: \"kubernetes.io/projected/364be77d-ab70-49ff-9fa4-2b42f49b718b-kube-api-access-sn25n\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566797 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/529c3d09-735a-4157-b35b-0adb74424387-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566814 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppp6k\" (UniqueName: \"kubernetes.io/projected/54366321-671a-46e6-a4ca-a834ba087059-kube-api-access-ppp6k\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566838 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5abee0e6-194a-4457-86dc-6d27a745589b-config\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566855 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-config\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566874 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-serving-cert\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566890 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-serving-cert\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566920 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-trusted-ca\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566938 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5j4h\" (UniqueName: \"kubernetes.io/projected/5abee0e6-194a-4457-86dc-6d27a745589b-kube-api-access-p5j4h\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566956 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b28c51f3-e546-4bee-923f-e27b8c7c7567-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566972 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-audit-policies\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.566987 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db8zr\" (UniqueName: \"kubernetes.io/projected/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-kube-api-access-db8zr\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567004 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567021 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567037 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-service-ca\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567056 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5abee0e6-194a-4457-86dc-6d27a745589b-images\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567073 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a761b6de-3f12-4781-9a64-50bc79d6f482-metrics-tls\") pod \"dns-operator-744455d44c-mbn6w\" (UID: \"a761b6de-3f12-4781-9a64-50bc79d6f482\") " pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567093 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7758de03-0df5-45b4-8b7b-2a071740b2e0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567113 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjpg\" (UniqueName: \"kubernetes.io/projected/a3ed3536-f707-47a5-b3df-3ec72bec773b-kube-api-access-vnjpg\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567131 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-service-ca\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567147 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567169 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-serving-cert\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567186 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567201 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqwgh\" (UniqueName: \"kubernetes.io/projected/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-kube-api-access-mqwgh\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567216 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-encryption-config\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567232 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-ca\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567248 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-client-ca\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567271 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/529c3d09-735a-4157-b35b-0adb74424387-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567294 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c414497-e825-484c-bfb1-1e9dc5e4d9d2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z4722\" (UID: \"8c414497-e825-484c-bfb1-1e9dc5e4d9d2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567315 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t6cm\" (UniqueName: \"kubernetes.io/projected/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-kube-api-access-8t6cm\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567331 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-config\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567349 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-serving-cert\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567364 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567383 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567399 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567419 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567435 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-image-import-ca\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567453 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-audit\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567471 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lgr5\" (UniqueName: \"kubernetes.io/projected/2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a-kube-api-access-9lgr5\") pod \"multus-admission-controller-857f4d67dd-6gdcn\" (UID: \"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567489 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-oauth-serving-cert\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567504 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-oauth-config\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567519 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-etcd-client\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567536 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-etcd-serving-ca\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567551 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-audit-dir\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567570 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-config\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567586 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b28c51f3-e546-4bee-923f-e27b8c7c7567-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567609 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567624 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4cdf1f3-c551-4561-b731-a5413a13a724-trusted-ca\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567640 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d163c22-5dce-4cae-b536-c86dabcf6a91-serving-cert\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567656 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567671 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-policies\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567686 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvbbc\" (UniqueName: \"kubernetes.io/projected/8c414497-e825-484c-bfb1-1e9dc5e4d9d2-kube-api-access-vvbbc\") pod \"cluster-samples-operator-665b6dd947-z4722\" (UID: \"8c414497-e825-484c-bfb1-1e9dc5e4d9d2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.567711 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-trusted-ca-bundle\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.568908 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-c5n5d"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.570626 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-config\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.570915 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-auth-proxy-config\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.571061 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.571408 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-client-ca\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.571861 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-config\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.572153 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5abee0e6-194a-4457-86dc-6d27a745589b-config\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.572356 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b28c51f3-e546-4bee-923f-e27b8c7c7567-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.572868 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-service-ca-bundle\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.573010 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-client-ca\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.573863 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.573887 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-q72r8"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.573919 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.573931 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.574026 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.574102 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-trusted-ca-bundle\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.574292 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t72fg"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.574475 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-config\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.574959 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-oauth-serving-cert\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.575230 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-service-ca\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.575867 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5abee0e6-194a-4457-86dc-6d27a745589b-images\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.575979 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ce1ea1-c409-4850-855b-fa8316d0d5e6-serving-cert\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.576121 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-config\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.576769 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-config\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.576791 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32ce1ea1-c409-4850-855b-fa8316d0d5e6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.576944 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-oauth-config\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.578314 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.578930 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e95bef62-5bd2-4a64-b1c5-df780f79ac51-console-serving-cert\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.579553 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-serving-cert\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.580230 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lbqfj"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.580869 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b28c51f3-e546-4bee-923f-e27b8c7c7567-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.580905 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-machine-approver-tls\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.581312 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.581550 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5abee0e6-194a-4457-86dc-6d27a745589b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.581600 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vw7pc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.582869 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.585353 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-b6wlp"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.586084 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3ed3536-f707-47a5-b3df-3ec72bec773b-serving-cert\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.586815 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mbn6w"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.588001 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.589287 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.590432 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t56br"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.591652 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.593108 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9r9nf"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.594080 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mz42v"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.595601 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.597611 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.599136 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.600100 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.601073 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c5n5d"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.601740 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.602274 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.603408 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.604835 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6gdcn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.606097 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.608581 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.608603 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9k26f"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.609241 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.610405 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-w7whn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.611612 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-465wh"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.612029 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.612226 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.612534 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.613468 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.614530 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-k4phl"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.615643 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.616946 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-w7whn"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.618003 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcw2q"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.619805 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jk8m2"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.620793 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jk8m2"] Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.620889 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.635704 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.641865 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.662305 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668375 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k77m6\" (UniqueName: \"kubernetes.io/projected/7d163c22-5dce-4cae-b536-c86dabcf6a91-kube-api-access-k77m6\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668421 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668445 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-node-pullsecrets\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668463 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54366321-671a-46e6-a4ca-a834ba087059-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668483 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668501 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4cdf1f3-c551-4561-b731-a5413a13a724-metrics-tls\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc49h\" (UniqueName: \"kubernetes.io/projected/529c3d09-735a-4157-b35b-0adb74424387-kube-api-access-lc49h\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668548 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/529c3d09-735a-4157-b35b-0adb74424387-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668566 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-config\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668587 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-encryption-config\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668607 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7758de03-0df5-45b4-8b7b-2a071740b2e0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668652 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668680 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-config\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668700 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnrvv\" (UniqueName: \"kubernetes.io/projected/c207de79-b8c1-47f5-8e4b-5f320d760588-kube-api-access-pnrvv\") pod \"downloads-7954f5f757-9r9nf\" (UID: \"c207de79-b8c1-47f5-8e4b-5f320d760588\") " pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668717 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4cdf1f3-c551-4561-b731-a5413a13a724-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668735 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668753 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-config\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668776 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-dir\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668793 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668810 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-etcd-client\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668829 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668849 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668884 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6gdcn\" (UID: \"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668927 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54366321-671a-46e6-a4ca-a834ba087059-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668946 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668964 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nq9x\" (UniqueName: \"kubernetes.io/projected/e4cdf1f3-c551-4561-b731-a5413a13a724-kube-api-access-6nq9x\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.668983 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669003 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669023 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x478z\" (UniqueName: \"kubernetes.io/projected/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-kube-api-access-x478z\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669043 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-client\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669060 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-serving-cert\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669087 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-audit-dir\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669106 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn25n\" (UniqueName: \"kubernetes.io/projected/364be77d-ab70-49ff-9fa4-2b42f49b718b-kube-api-access-sn25n\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669127 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/529c3d09-735a-4157-b35b-0adb74424387-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669146 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppp6k\" (UniqueName: \"kubernetes.io/projected/54366321-671a-46e6-a4ca-a834ba087059-kube-api-access-ppp6k\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669172 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-serving-cert\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669191 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-serving-cert\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669208 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-trusted-ca\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669225 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-audit-policies\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669244 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db8zr\" (UniqueName: \"kubernetes.io/projected/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-kube-api-access-db8zr\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669267 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669287 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669306 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a761b6de-3f12-4781-9a64-50bc79d6f482-metrics-tls\") pod \"dns-operator-744455d44c-mbn6w\" (UID: \"a761b6de-3f12-4781-9a64-50bc79d6f482\") " pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669325 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7758de03-0df5-45b4-8b7b-2a071740b2e0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669351 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-service-ca\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669368 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669386 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-serving-cert\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669404 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqwgh\" (UniqueName: \"kubernetes.io/projected/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-kube-api-access-mqwgh\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669422 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669439 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-encryption-config\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669455 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-ca\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669475 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/529c3d09-735a-4157-b35b-0adb74424387-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669493 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c414497-e825-484c-bfb1-1e9dc5e4d9d2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z4722\" (UID: \"8c414497-e825-484c-bfb1-1e9dc5e4d9d2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669512 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-config\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669529 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t6cm\" (UniqueName: \"kubernetes.io/projected/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-kube-api-access-8t6cm\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669557 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669584 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669613 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669637 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-image-import-ca\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669657 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-audit\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669677 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lgr5\" (UniqueName: \"kubernetes.io/projected/2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a-kube-api-access-9lgr5\") pod \"multus-admission-controller-857f4d67dd-6gdcn\" (UID: \"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669694 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-etcd-client\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669711 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-etcd-serving-ca\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669728 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-audit-dir\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669759 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669776 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4cdf1f3-c551-4561-b731-a5413a13a724-trusted-ca\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669794 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d163c22-5dce-4cae-b536-c86dabcf6a91-serving-cert\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669813 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669831 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-policies\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669849 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvbbc\" (UniqueName: \"kubernetes.io/projected/8c414497-e825-484c-bfb1-1e9dc5e4d9d2-kube-api-access-vvbbc\") pod \"cluster-samples-operator-665b6dd947-z4722\" (UID: \"8c414497-e825-484c-bfb1-1e9dc5e4d9d2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669868 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78d2w\" (UniqueName: \"kubernetes.io/projected/a761b6de-3f12-4781-9a64-50bc79d6f482-kube-api-access-78d2w\") pod \"dns-operator-744455d44c-mbn6w\" (UID: \"a761b6de-3f12-4781-9a64-50bc79d6f482\") " pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669888 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtckm\" (UniqueName: \"kubernetes.io/projected/7758de03-0df5-45b4-8b7b-2a071740b2e0-kube-api-access-jtckm\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.669971 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.670680 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.671036 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-node-pullsecrets\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.671582 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-config\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.672128 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-config\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.672797 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.672827 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-ca\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.672799 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/529c3d09-735a-4157-b35b-0adb74424387-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.673199 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.673215 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.672800 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-config\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.673568 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54366321-671a-46e6-a4ca-a834ba087059-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.673836 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-audit-policies\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.674570 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4cdf1f3-c551-4561-b731-a5413a13a724-metrics-tls\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.674834 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/529c3d09-735a-4157-b35b-0adb74424387-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.674927 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-encryption-config\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.675393 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-audit-dir\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.675464 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-dir\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.675580 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-etcd-serving-ca\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.675651 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-audit-dir\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.676759 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-encryption-config\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.676777 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.677001 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-trusted-ca\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.677107 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4cdf1f3-c551-4561-b731-a5413a13a724-trusted-ca\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.677843 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.677880 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54366321-671a-46e6-a4ca-a834ba087059-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.678204 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.678442 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-policies\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.678551 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.678724 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-audit\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.679356 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-client\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.679472 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-etcd-client\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.679793 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-serving-cert\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.679810 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-serving-cert\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.680040 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.680328 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.680338 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.680941 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.681024 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-serving-cert\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.681396 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.681764 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.681968 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-etcd-client\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.682779 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-serving-cert\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.683116 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.683325 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.683559 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c414497-e825-484c-bfb1-1e9dc5e4d9d2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z4722\" (UID: \"8c414497-e825-484c-bfb1-1e9dc5e4d9d2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.683780 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d163c22-5dce-4cae-b536-c86dabcf6a91-serving-cert\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.684264 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-image-import-ca\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.703708 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.721200 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.741241 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.761026 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.772285 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a761b6de-3f12-4781-9a64-50bc79d6f482-metrics-tls\") pod \"dns-operator-744455d44c-mbn6w\" (UID: \"a761b6de-3f12-4781-9a64-50bc79d6f482\") " pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.782224 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.801843 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.809506 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.823283 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.843099 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.846303 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-config\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.861367 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.882835 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.902341 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.907338 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d163c22-5dce-4cae-b536-c86dabcf6a91-etcd-service-ca\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.922244 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.943608 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.961082 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.970505 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7758de03-0df5-45b4-8b7b-2a071740b2e0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:24 crc kubenswrapper[4702]: I1007 12:26:24.981232 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.001416 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.006342 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7758de03-0df5-45b4-8b7b-2a071740b2e0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.023405 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.031799 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.042973 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.061354 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.068748 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.082497 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.100811 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.106623 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6gdcn\" (UID: \"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.122177 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.162317 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.182632 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.201718 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.221323 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.243257 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.263357 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.281873 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.303091 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.322894 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.341592 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.361871 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.381333 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.401469 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.422243 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.441430 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.463337 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.479677 4702 request.go:700] Waited for 1.009893083s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-sa-dockercfg-5xfcg&limit=500&resourceVersion=0 Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.482383 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.501675 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.542176 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.561775 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.582481 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.601686 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.622800 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.641823 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.661659 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.683388 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.702310 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.723054 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.743021 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.762172 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.782051 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.801595 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.828216 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.842350 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.861888 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.881700 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.901976 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.922565 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.941559 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.962294 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 12:26:25 crc kubenswrapper[4702]: I1007 12:26:25.983077 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.001941 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.020955 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.041695 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.061687 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.082850 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.119425 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlsc7\" (UniqueName: \"kubernetes.io/projected/f8056a9f-96dd-4c77-902d-0862fbc3b1ec-kube-api-access-qlsc7\") pod \"machine-approver-56656f9798-hwfzn\" (UID: \"f8056a9f-96dd-4c77-902d-0862fbc3b1ec\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.137779 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s258v\" (UniqueName: \"kubernetes.io/projected/b28c51f3-e546-4bee-923f-e27b8c7c7567-kube-api-access-s258v\") pod \"openshift-controller-manager-operator-756b6f6bc6-5bbhq\" (UID: \"b28c51f3-e546-4bee-923f-e27b8c7c7567\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.166540 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl78l\" (UniqueName: \"kubernetes.io/projected/e95bef62-5bd2-4a64-b1c5-df780f79ac51-kube-api-access-hl78l\") pod \"console-f9d7485db-q72r8\" (UID: \"e95bef62-5bd2-4a64-b1c5-df780f79ac51\") " pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.179569 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5j4h\" (UniqueName: \"kubernetes.io/projected/5abee0e6-194a-4457-86dc-6d27a745589b-kube-api-access-p5j4h\") pod \"machine-api-operator-5694c8668f-qtvlc\" (UID: \"5abee0e6-194a-4457-86dc-6d27a745589b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.197519 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqpk8\" (UniqueName: \"kubernetes.io/projected/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-kube-api-access-lqpk8\") pod \"route-controller-manager-6576b87f9c-5dwsj\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.199180 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.201340 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.222014 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.243142 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.272288 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.281569 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.285165 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjpg\" (UniqueName: \"kubernetes.io/projected/a3ed3536-f707-47a5-b3df-3ec72bec773b-kube-api-access-vnjpg\") pod \"controller-manager-879f6c89f-n9lq9\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.290988 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.302631 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.310375 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf2rd\" (UniqueName: \"kubernetes.io/projected/32ce1ea1-c409-4850-855b-fa8316d0d5e6-kube-api-access-rf2rd\") pod \"authentication-operator-69f744f599-b6wlp\" (UID: \"32ce1ea1-c409-4850-855b-fa8316d0d5e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.323171 4702 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.342385 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.344573 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.362328 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.382534 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.405864 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.425535 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.432447 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.449691 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq"] Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.449970 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.464278 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.464296 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 07 12:26:26 crc kubenswrapper[4702]: W1007 12:26:26.474603 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb28c51f3_e546_4bee_923f_e27b8c7c7567.slice/crio-a397e5867bb8b519d22c348ed3179637f353c41418b15149e646fe141eac0806 WatchSource:0}: Error finding container a397e5867bb8b519d22c348ed3179637f353c41418b15149e646fe141eac0806: Status 404 returned error can't find the container with id a397e5867bb8b519d22c348ed3179637f353c41418b15149e646fe141eac0806 Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.478488 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" event={"ID":"f8056a9f-96dd-4c77-902d-0862fbc3b1ec","Type":"ContainerStarted","Data":"7fad112e90e036188bc3fe13435c67ade3901be72820023fe34276c4a2fb0ca6"} Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.485134 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.501666 4702 request.go:700] Waited for 1.832951447s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/serviceaccounts/etcd-operator/token Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.506203 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-q72r8"] Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.520037 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k77m6\" (UniqueName: \"kubernetes.io/projected/7d163c22-5dce-4cae-b536-c86dabcf6a91-kube-api-access-k77m6\") pod \"etcd-operator-b45778765-t72fg\" (UID: \"7d163c22-5dce-4cae-b536-c86dabcf6a91\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:26 crc kubenswrapper[4702]: W1007 12:26:26.522968 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode95bef62_5bd2_4a64_b1c5_df780f79ac51.slice/crio-bb58b3aaa7374f7e480c4ecbb87863b8090e3ec0b39d0cbc018649a3c46e1499 WatchSource:0}: Error finding container bb58b3aaa7374f7e480c4ecbb87863b8090e3ec0b39d0cbc018649a3c46e1499: Status 404 returned error can't find the container with id bb58b3aaa7374f7e480c4ecbb87863b8090e3ec0b39d0cbc018649a3c46e1499 Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.548161 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nq9x\" (UniqueName: \"kubernetes.io/projected/e4cdf1f3-c551-4561-b731-a5413a13a724-kube-api-access-6nq9x\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.560806 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc49h\" (UniqueName: \"kubernetes.io/projected/529c3d09-735a-4157-b35b-0adb74424387-kube-api-access-lc49h\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.561236 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qtvlc"] Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.577427 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lgr5\" (UniqueName: \"kubernetes.io/projected/2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a-kube-api-access-9lgr5\") pod \"multus-admission-controller-857f4d67dd-6gdcn\" (UID: \"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:26 crc kubenswrapper[4702]: W1007 12:26:26.589457 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5abee0e6_194a_4457_86dc_6d27a745589b.slice/crio-63162c84d3ab2a49406008c34e48a9420b8bd0e0ab4c5b9abd79dcdf822edb2b WatchSource:0}: Error finding container 63162c84d3ab2a49406008c34e48a9420b8bd0e0ab4c5b9abd79dcdf822edb2b: Status 404 returned error can't find the container with id 63162c84d3ab2a49406008c34e48a9420b8bd0e0ab4c5b9abd79dcdf822edb2b Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.603097 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnrvv\" (UniqueName: \"kubernetes.io/projected/c207de79-b8c1-47f5-8e4b-5f320d760588-kube-api-access-pnrvv\") pod \"downloads-7954f5f757-9r9nf\" (UID: \"c207de79-b8c1-47f5-8e4b-5f320d760588\") " pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.620479 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj"] Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.626357 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4cdf1f3-c551-4561-b731-a5413a13a724-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2mpvg\" (UID: \"e4cdf1f3-c551-4561-b731-a5413a13a724\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.636300 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/529c3d09-735a-4157-b35b-0adb74424387-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-78mw8\" (UID: \"529c3d09-735a-4157-b35b-0adb74424387\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.661777 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.667156 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn25n\" (UniqueName: \"kubernetes.io/projected/364be77d-ab70-49ff-9fa4-2b42f49b718b-kube-api-access-sn25n\") pod \"oauth-openshift-558db77b4-lbqfj\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.692235 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x478z\" (UniqueName: \"kubernetes.io/projected/e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd-kube-api-access-x478z\") pod \"apiserver-76f77b778f-mz42v\" (UID: \"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd\") " pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.698984 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppp6k\" (UniqueName: \"kubernetes.io/projected/54366321-671a-46e6-a4ca-a834ba087059-kube-api-access-ppp6k\") pod \"openshift-apiserver-operator-796bbdcf4f-j8v4k\" (UID: \"54366321-671a-46e6-a4ca-a834ba087059\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.709838 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.721620 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-b6wlp"] Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.731045 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:26 crc kubenswrapper[4702]: W1007 12:26:26.736649 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32ce1ea1_c409_4850_855b_fa8316d0d5e6.slice/crio-34e66d36b50e80b9061f592a4decfd9c7d4f444bcc9ef0bd4c8f32e06eb73c1b WatchSource:0}: Error finding container 34e66d36b50e80b9061f592a4decfd9c7d4f444bcc9ef0bd4c8f32e06eb73c1b: Status 404 returned error can't find the container with id 34e66d36b50e80b9061f592a4decfd9c7d4f444bcc9ef0bd4c8f32e06eb73c1b Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.737442 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.738802 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t6cm\" (UniqueName: \"kubernetes.io/projected/b7763d6d-7a4e-4aa3-80c0-fc4e27a66443-kube-api-access-8t6cm\") pod \"openshift-config-operator-7777fb866f-9k26f\" (UID: \"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.738826 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db8zr\" (UniqueName: \"kubernetes.io/projected/91bd4d74-eea0-42d9-9cb6-52c062a2b04c-kube-api-access-db8zr\") pod \"apiserver-7bbb656c7d-6vxk4\" (UID: \"91bd4d74-eea0-42d9-9cb6-52c062a2b04c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.747356 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.755359 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.760257 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eb394c2-b11a-4466-a9aa-708bbb4ac4e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf7kg\" (UID: \"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.763583 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.770091 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.785983 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqwgh\" (UniqueName: \"kubernetes.io/projected/dc1f4290-b2cb-469b-9fa2-94f7ab35f42b-kube-api-access-mqwgh\") pod \"console-operator-58897d9998-t56br\" (UID: \"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b\") " pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.801339 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5ccb5e91-d8bc-4272-8656-d05e57e1d79c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-q5qbc\" (UID: \"5ccb5e91-d8bc-4272-8656-d05e57e1d79c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.823279 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvbbc\" (UniqueName: \"kubernetes.io/projected/8c414497-e825-484c-bfb1-1e9dc5e4d9d2-kube-api-access-vvbbc\") pod \"cluster-samples-operator-665b6dd947-z4722\" (UID: \"8c414497-e825-484c-bfb1-1e9dc5e4d9d2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.825678 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.836001 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtckm\" (UniqueName: \"kubernetes.io/projected/7758de03-0df5-45b4-8b7b-2a071740b2e0-kube-api-access-jtckm\") pod \"kube-storage-version-migrator-operator-b67b599dd-n5p5n\" (UID: \"7758de03-0df5-45b4-8b7b-2a071740b2e0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.840345 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.848813 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.855055 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.861539 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78d2w\" (UniqueName: \"kubernetes.io/projected/a761b6de-3f12-4781-9a64-50bc79d6f482-kube-api-access-78d2w\") pod \"dns-operator-744455d44c-mbn6w\" (UID: \"a761b6de-3f12-4781-9a64-50bc79d6f482\") " pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.864510 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8"] Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.908486 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-bound-sa-token\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909183 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-tls\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909209 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-trusted-ca\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909233 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909254 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-certificates\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909288 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e0a452b7-fe35-404d-8864-ff8d6be236da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909315 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz5bw\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-kube-api-access-qz5bw\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.909346 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e0a452b7-fe35-404d-8864-ff8d6be236da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:26 crc kubenswrapper[4702]: E1007 12:26:26.909755 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.409739465 +0000 UTC m=+143.585475581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.980238 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" Oct 07 12:26:26 crc kubenswrapper[4702]: I1007 12:26:26.987170 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9lq9"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.010584 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.010912 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.510865518 +0000 UTC m=+143.686601634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.010952 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtvgv\" (UniqueName: \"kubernetes.io/projected/802177f9-2e2e-4cb8-895f-686ed017c9e3-kube-api-access-wtvgv\") pod \"control-plane-machine-set-operator-78cbb6b69f-47q4n\" (UID: \"802177f9-2e2e-4cb8-895f-686ed017c9e3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011015 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/183df67a-7508-4b77-ac96-ec05258e68ee-node-bootstrap-token\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011046 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwhh\" (UniqueName: \"kubernetes.io/projected/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-kube-api-access-wnwhh\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011106 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbqzh\" (UniqueName: \"kubernetes.io/projected/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-kube-api-access-rbqzh\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011145 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e0a452b7-fe35-404d-8864-ff8d6be236da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011177 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t5sh\" (UniqueName: \"kubernetes.io/projected/9596f202-95f9-45ae-92be-6162cc42a574-kube-api-access-7t5sh\") pod \"migrator-59844c95c7-gnc24\" (UID: \"9596f202-95f9-45ae-92be-6162cc42a574\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011193 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-metrics-certs\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011238 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz5bw\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-kube-api-access-qz5bw\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011257 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/48f7df49-3241-4dd4-b579-ac0d7436f9ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011284 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24xtq\" (UniqueName: \"kubernetes.io/projected/56bcfcc7-401b-4261-b7de-2a28515c6267-kube-api-access-24xtq\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011314 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-apiservice-cert\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011340 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/56bcfcc7-401b-4261-b7de-2a28515c6267-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011357 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-config-volume\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011372 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x8pt\" (UniqueName: \"kubernetes.io/projected/183df67a-7508-4b77-ac96-ec05258e68ee-kube-api-access-5x8pt\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011406 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6qzn\" (UniqueName: \"kubernetes.io/projected/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-kube-api-access-j6qzn\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011424 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-socket-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011441 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e0a452b7-fe35-404d-8864-ff8d6be236da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011457 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/56bcfcc7-401b-4261-b7de-2a28515c6267-images\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011477 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr2ft\" (UniqueName: \"kubernetes.io/projected/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-kube-api-access-xr2ft\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011495 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529d8a79-e1d3-4001-b791-4c440487bad9-config\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011594 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2rm\" (UniqueName: \"kubernetes.io/projected/529d8a79-e1d3-4001-b791-4c440487bad9-kube-api-access-9t2rm\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011612 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-bound-sa-token\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011662 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011720 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jngnc\" (UniqueName: \"kubernetes.io/projected/6f164dd1-c728-4d35-b7f8-12fca172497d-kube-api-access-jngnc\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011746 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/48f7df49-3241-4dd4-b579-ac0d7436f9ab-srv-cert\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011761 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2a92879d-06fd-4c7f-853c-360e5215a574-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.011807 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9ca3c868-88b7-49e3-acd0-ba172b13b24f-signing-key\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012016 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdqzv\" (UniqueName: \"kubernetes.io/projected/48f7df49-3241-4dd4-b579-ac0d7436f9ab-kube-api-access-pdqzv\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012062 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-tls\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012077 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012129 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-metrics-tls\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012222 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-certificates\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012238 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5101fb34-4898-4acc-916d-3e2d69a46c1f-cert\") pod \"ingress-canary-jk8m2\" (UID: \"5101fb34-4898-4acc-916d-3e2d69a46c1f\") " pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012263 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/56bcfcc7-401b-4261-b7de-2a28515c6267-proxy-tls\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012319 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frpsj\" (UniqueName: \"kubernetes.io/projected/5101fb34-4898-4acc-916d-3e2d69a46c1f-kube-api-access-frpsj\") pod \"ingress-canary-jk8m2\" (UID: \"5101fb34-4898-4acc-916d-3e2d69a46c1f\") " pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012335 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-plugins-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012364 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbx9f\" (UniqueName: \"kubernetes.io/projected/bf6cffa0-4366-4806-a956-34437988bdcf-kube-api-access-gbx9f\") pod \"package-server-manager-789f6589d5-l9r92\" (UID: \"bf6cffa0-4366-4806-a956-34437988bdcf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012397 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-default-certificate\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012485 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012504 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012533 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2a92879d-06fd-4c7f-853c-360e5215a574-srv-cert\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012563 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5da410ed-29be-4a38-b486-89f0dd0252ae-config-volume\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012588 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-config\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012602 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-registration-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012628 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf6cffa0-4366-4806-a956-34437988bdcf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l9r92\" (UID: \"bf6cffa0-4366-4806-a956-34437988bdcf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012676 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3602b00e-3c66-4dd4-94fb-cc4421ea7566-service-ca-bundle\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012723 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-tmpfs\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012749 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/802177f9-2e2e-4cb8-895f-686ed017c9e3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-47q4n\" (UID: \"802177f9-2e2e-4cb8-895f-686ed017c9e3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012776 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-proxy-tls\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012800 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r7jm\" (UniqueName: \"kubernetes.io/projected/2a92879d-06fd-4c7f-853c-360e5215a574-kube-api-access-8r7jm\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012846 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/183df67a-7508-4b77-ac96-ec05258e68ee-certs\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012872 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9ca3c868-88b7-49e3-acd0-ba172b13b24f-signing-cabundle\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012949 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012967 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhg52\" (UniqueName: \"kubernetes.io/projected/3602b00e-3c66-4dd4-94fb-cc4421ea7566-kube-api-access-vhg52\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.012993 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-webhook-cert\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013016 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-trusted-ca\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013039 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt47q\" (UniqueName: \"kubernetes.io/projected/5da410ed-29be-4a38-b486-89f0dd0252ae-kube-api-access-rt47q\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013063 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-stats-auth\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013083 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/529d8a79-e1d3-4001-b791-4c440487bad9-serving-cert\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013123 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-csi-data-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013148 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjd57\" (UniqueName: \"kubernetes.io/projected/9ca3c868-88b7-49e3-acd0-ba172b13b24f-kube-api-access-sjd57\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013164 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5da410ed-29be-4a38-b486-89f0dd0252ae-secret-volume\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013200 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013222 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-mountpoint-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.013718 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e0a452b7-fe35-404d-8864-ff8d6be236da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.022825 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.023579 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.523556463 +0000 UTC m=+143.699292779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.026073 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-certificates\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.027388 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.027399 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-trusted-ca\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.027971 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e0a452b7-fe35-404d-8864-ff8d6be236da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.034425 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-tls\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.052761 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.062929 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz5bw\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-kube-api-access-qz5bw\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.084647 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-bound-sa-token\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116039 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116270 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5101fb34-4898-4acc-916d-3e2d69a46c1f-cert\") pod \"ingress-canary-jk8m2\" (UID: \"5101fb34-4898-4acc-916d-3e2d69a46c1f\") " pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116297 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/56bcfcc7-401b-4261-b7de-2a28515c6267-proxy-tls\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116324 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbx9f\" (UniqueName: \"kubernetes.io/projected/bf6cffa0-4366-4806-a956-34437988bdcf-kube-api-access-gbx9f\") pod \"package-server-manager-789f6589d5-l9r92\" (UID: \"bf6cffa0-4366-4806-a956-34437988bdcf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116344 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frpsj\" (UniqueName: \"kubernetes.io/projected/5101fb34-4898-4acc-916d-3e2d69a46c1f-kube-api-access-frpsj\") pod \"ingress-canary-jk8m2\" (UID: \"5101fb34-4898-4acc-916d-3e2d69a46c1f\") " pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116362 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-plugins-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116395 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-default-certificate\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116416 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116433 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116450 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2a92879d-06fd-4c7f-853c-360e5215a574-srv-cert\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116469 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5da410ed-29be-4a38-b486-89f0dd0252ae-config-volume\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116489 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-config\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116506 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-registration-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116523 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf6cffa0-4366-4806-a956-34437988bdcf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l9r92\" (UID: \"bf6cffa0-4366-4806-a956-34437988bdcf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116542 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3602b00e-3c66-4dd4-94fb-cc4421ea7566-service-ca-bundle\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116564 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-tmpfs\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116583 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/802177f9-2e2e-4cb8-895f-686ed017c9e3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-47q4n\" (UID: \"802177f9-2e2e-4cb8-895f-686ed017c9e3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116601 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-proxy-tls\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116621 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r7jm\" (UniqueName: \"kubernetes.io/projected/2a92879d-06fd-4c7f-853c-360e5215a574-kube-api-access-8r7jm\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116639 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/183df67a-7508-4b77-ac96-ec05258e68ee-certs\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116669 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9ca3c868-88b7-49e3-acd0-ba172b13b24f-signing-cabundle\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116688 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116707 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhg52\" (UniqueName: \"kubernetes.io/projected/3602b00e-3c66-4dd4-94fb-cc4421ea7566-kube-api-access-vhg52\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116728 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-webhook-cert\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116747 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt47q\" (UniqueName: \"kubernetes.io/projected/5da410ed-29be-4a38-b486-89f0dd0252ae-kube-api-access-rt47q\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116766 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-stats-auth\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116785 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/529d8a79-e1d3-4001-b791-4c440487bad9-serving-cert\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116806 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjd57\" (UniqueName: \"kubernetes.io/projected/9ca3c868-88b7-49e3-acd0-ba172b13b24f-kube-api-access-sjd57\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116823 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5da410ed-29be-4a38-b486-89f0dd0252ae-secret-volume\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116843 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-csi-data-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116870 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-mountpoint-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.116911 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtvgv\" (UniqueName: \"kubernetes.io/projected/802177f9-2e2e-4cb8-895f-686ed017c9e3-kube-api-access-wtvgv\") pod \"control-plane-machine-set-operator-78cbb6b69f-47q4n\" (UID: \"802177f9-2e2e-4cb8-895f-686ed017c9e3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.117587 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-plugins-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118393 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/183df67a-7508-4b77-ac96-ec05258e68ee-node-bootstrap-token\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118425 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-config\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118442 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwhh\" (UniqueName: \"kubernetes.io/projected/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-kube-api-access-wnwhh\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.118614 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.61859279 +0000 UTC m=+143.794328906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118681 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-csi-data-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118775 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbqzh\" (UniqueName: \"kubernetes.io/projected/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-kube-api-access-rbqzh\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118800 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t5sh\" (UniqueName: \"kubernetes.io/projected/9596f202-95f9-45ae-92be-6162cc42a574-kube-api-access-7t5sh\") pod \"migrator-59844c95c7-gnc24\" (UID: \"9596f202-95f9-45ae-92be-6162cc42a574\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118822 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-metrics-certs\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.118970 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3602b00e-3c66-4dd4-94fb-cc4421ea7566-service-ca-bundle\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.119096 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-registration-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.120677 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-mountpoint-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.121161 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.121846 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-tmpfs\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.121941 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9ca3c868-88b7-49e3-acd0-ba172b13b24f-signing-cabundle\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.122491 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.122643 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/48f7df49-3241-4dd4-b579-ac0d7436f9ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.122789 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24xtq\" (UniqueName: \"kubernetes.io/projected/56bcfcc7-401b-4261-b7de-2a28515c6267-kube-api-access-24xtq\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.122817 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5da410ed-29be-4a38-b486-89f0dd0252ae-config-volume\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.123374 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-apiservice-cert\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.125453 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/56bcfcc7-401b-4261-b7de-2a28515c6267-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.125567 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-config-volume\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.125699 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x8pt\" (UniqueName: \"kubernetes.io/projected/183df67a-7508-4b77-ac96-ec05258e68ee-kube-api-access-5x8pt\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.125793 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6qzn\" (UniqueName: \"kubernetes.io/projected/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-kube-api-access-j6qzn\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.126480 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-config-volume\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.129030 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/56bcfcc7-401b-4261-b7de-2a28515c6267-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.129292 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/56bcfcc7-401b-4261-b7de-2a28515c6267-images\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.129819 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/56bcfcc7-401b-4261-b7de-2a28515c6267-images\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.130347 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/56bcfcc7-401b-4261-b7de-2a28515c6267-proxy-tls\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.134461 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr2ft\" (UniqueName: \"kubernetes.io/projected/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-kube-api-access-xr2ft\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.136669 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-socket-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.135794 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.134951 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/529d8a79-e1d3-4001-b791-4c440487bad9-serving-cert\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.136831 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-socket-dir\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.136869 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529d8a79-e1d3-4001-b791-4c440487bad9-config\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137047 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2rm\" (UniqueName: \"kubernetes.io/projected/529d8a79-e1d3-4001-b791-4c440487bad9-kube-api-access-9t2rm\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137081 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137111 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jngnc\" (UniqueName: \"kubernetes.io/projected/6f164dd1-c728-4d35-b7f8-12fca172497d-kube-api-access-jngnc\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137138 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/48f7df49-3241-4dd4-b579-ac0d7436f9ab-srv-cert\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137156 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2a92879d-06fd-4c7f-853c-360e5215a574-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137182 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9ca3c868-88b7-49e3-acd0-ba172b13b24f-signing-key\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137244 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdqzv\" (UniqueName: \"kubernetes.io/projected/48f7df49-3241-4dd4-b579-ac0d7436f9ab-kube-api-access-pdqzv\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137312 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137335 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-metrics-tls\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.137576 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/183df67a-7508-4b77-ac96-ec05258e68ee-certs\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.139069 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-webhook-cert\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.140289 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-default-certificate\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.142038 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-apiservice-cert\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.142276 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529d8a79-e1d3-4001-b791-4c440487bad9-config\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.143769 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/183df67a-7508-4b77-ac96-ec05258e68ee-node-bootstrap-token\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.143962 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/802177f9-2e2e-4cb8-895f-686ed017c9e3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-47q4n\" (UID: \"802177f9-2e2e-4cb8-895f-686ed017c9e3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.144945 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-stats-auth\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.144967 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-proxy-tls\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.145500 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/48f7df49-3241-4dd4-b579-ac0d7436f9ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.147999 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf6cffa0-4366-4806-a956-34437988bdcf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l9r92\" (UID: \"bf6cffa0-4366-4806-a956-34437988bdcf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.148450 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2a92879d-06fd-4c7f-853c-360e5215a574-srv-cert\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.148784 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5da410ed-29be-4a38-b486-89f0dd0252ae-secret-volume\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.148936 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3602b00e-3c66-4dd4-94fb-cc4421ea7566-metrics-certs\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.149058 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5101fb34-4898-4acc-916d-3e2d69a46c1f-cert\") pod \"ingress-canary-jk8m2\" (UID: \"5101fb34-4898-4acc-916d-3e2d69a46c1f\") " pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.155714 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-metrics-tls\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.163179 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9ca3c868-88b7-49e3-acd0-ba172b13b24f-signing-key\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.163206 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.163266 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/48f7df49-3241-4dd4-b579-ac0d7436f9ab-srv-cert\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.163299 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frpsj\" (UniqueName: \"kubernetes.io/projected/5101fb34-4898-4acc-916d-3e2d69a46c1f-kube-api-access-frpsj\") pod \"ingress-canary-jk8m2\" (UID: \"5101fb34-4898-4acc-916d-3e2d69a46c1f\") " pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.164117 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2a92879d-06fd-4c7f-853c-360e5215a574-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.166031 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.181748 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10b7a68d-a361-4fef-b4de-761c0cb9ed5b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-szgpc\" (UID: \"10b7a68d-a361-4fef-b4de-761c0cb9ed5b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.201664 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r7jm\" (UniqueName: \"kubernetes.io/projected/2a92879d-06fd-4c7f-853c-360e5215a574-kube-api-access-8r7jm\") pod \"olm-operator-6b444d44fb-cgb8d\" (UID: \"2a92879d-06fd-4c7f-853c-360e5215a574\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.211247 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.227838 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwhh\" (UniqueName: \"kubernetes.io/projected/d1c7c81c-7284-4c9c-96de-d7c546a8eb97-kube-api-access-wnwhh\") pod \"machine-config-controller-84d6567774-dcpzn\" (UID: \"d1c7c81c-7284-4c9c-96de-d7c546a8eb97\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.253883 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.256220 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mz42v"] Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.256470 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.75645603 +0000 UTC m=+143.932192146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.270831 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjd57\" (UniqueName: \"kubernetes.io/projected/9ca3c868-88b7-49e3-acd0-ba172b13b24f-kube-api-access-sjd57\") pod \"service-ca-9c57cc56f-k4phl\" (UID: \"9ca3c868-88b7-49e3-acd0-ba172b13b24f\") " pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.276059 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtvgv\" (UniqueName: \"kubernetes.io/projected/802177f9-2e2e-4cb8-895f-686ed017c9e3-kube-api-access-wtvgv\") pod \"control-plane-machine-set-operator-78cbb6b69f-47q4n\" (UID: \"802177f9-2e2e-4cb8-895f-686ed017c9e3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.287478 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbx9f\" (UniqueName: \"kubernetes.io/projected/bf6cffa0-4366-4806-a956-34437988bdcf-kube-api-access-gbx9f\") pod \"package-server-manager-789f6589d5-l9r92\" (UID: \"bf6cffa0-4366-4806-a956-34437988bdcf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.311427 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt47q\" (UniqueName: \"kubernetes.io/projected/5da410ed-29be-4a38-b486-89f0dd0252ae-kube-api-access-rt47q\") pod \"collect-profiles-29330655-552q2\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.322567 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jk8m2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.345271 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t5sh\" (UniqueName: \"kubernetes.io/projected/9596f202-95f9-45ae-92be-6162cc42a574-kube-api-access-7t5sh\") pod \"migrator-59844c95c7-gnc24\" (UID: \"9596f202-95f9-45ae-92be-6162cc42a574\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.349424 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24xtq\" (UniqueName: \"kubernetes.io/projected/56bcfcc7-401b-4261-b7de-2a28515c6267-kube-api-access-24xtq\") pod \"machine-config-operator-74547568cd-7lknr\" (UID: \"56bcfcc7-401b-4261-b7de-2a28515c6267\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.357153 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.357354 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.857315985 +0000 UTC m=+144.033052101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.357498 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.358168 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.858147299 +0000 UTC m=+144.033883415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.361345 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t72fg"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.366368 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9r9nf"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.381964 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbqzh\" (UniqueName: \"kubernetes.io/projected/cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06-kube-api-access-rbqzh\") pod \"packageserver-d55dfcdfc-2drjl\" (UID: \"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.386155 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhg52\" (UniqueName: \"kubernetes.io/projected/3602b00e-3c66-4dd4-94fb-cc4421ea7566-kube-api-access-vhg52\") pod \"router-default-5444994796-n66bc\" (UID: \"3602b00e-3c66-4dd4-94fb-cc4421ea7566\") " pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.408958 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x8pt\" (UniqueName: \"kubernetes.io/projected/183df67a-7508-4b77-ac96-ec05258e68ee-kube-api-access-5x8pt\") pod \"machine-config-server-465wh\" (UID: \"183df67a-7508-4b77-ac96-ec05258e68ee\") " pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.411312 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lbqfj"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.428685 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6qzn\" (UniqueName: \"kubernetes.io/projected/d8dd7d8a-f682-4492-8e4d-7318d345a9dd-kube-api-access-j6qzn\") pod \"csi-hostpathplugin-w7whn\" (UID: \"d8dd7d8a-f682-4492-8e4d-7318d345a9dd\") " pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.458799 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.459299 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.959070465 +0000 UTC m=+144.134806571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.459807 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.460324 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:27.960311271 +0000 UTC m=+144.136047387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.462695 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.471539 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.481685 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr2ft\" (UniqueName: \"kubernetes.io/projected/c336b672-f2b9-45ce-94ea-e56ed02c3b9f-kube-api-access-xr2ft\") pod \"dns-default-c5n5d\" (UID: \"c336b672-f2b9-45ce-94ea-e56ed02c3b9f\") " pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.483987 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jngnc\" (UniqueName: \"kubernetes.io/projected/6f164dd1-c728-4d35-b7f8-12fca172497d-kube-api-access-jngnc\") pod \"marketplace-operator-79b997595-pcw2q\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.494202 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.494325 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.494709 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.509803 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2rm\" (UniqueName: \"kubernetes.io/projected/529d8a79-e1d3-4001-b791-4c440487bad9-kube-api-access-9t2rm\") pod \"service-ca-operator-777779d784-5x5r2\" (UID: \"529d8a79-e1d3-4001-b791-4c440487bad9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.518154 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdqzv\" (UniqueName: \"kubernetes.io/projected/48f7df49-3241-4dd4-b579-ac0d7436f9ab-kube-api-access-pdqzv\") pod \"catalog-operator-68c6474976-b8hkt\" (UID: \"48f7df49-3241-4dd4-b579-ac0d7436f9ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.518495 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.527922 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.536971 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.538116 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" event={"ID":"529c3d09-735a-4157-b35b-0adb74424387","Type":"ContainerStarted","Data":"66644e5909afa3e3fff716bca469b550ead16ca212d195cc4f0b8f778321185d"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.538168 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" event={"ID":"529c3d09-735a-4157-b35b-0adb74424387","Type":"ContainerStarted","Data":"b6a52ff5a6f7314898469685cc2ab9648ebc9292ca15c50917051dc89ae56ad5"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.545859 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" event={"ID":"5abee0e6-194a-4457-86dc-6d27a745589b","Type":"ContainerStarted","Data":"fab01d5f2167934b1c099b4c7153021a598589a4e28877196fec94b03e626a2d"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.545927 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" event={"ID":"5abee0e6-194a-4457-86dc-6d27a745589b","Type":"ContainerStarted","Data":"6d975fb0ef2f23ac1dd798bf6a2f00486fd9f1ae1c4fff2c8248312f486f1cfd"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.545943 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" event={"ID":"5abee0e6-194a-4457-86dc-6d27a745589b","Type":"ContainerStarted","Data":"63162c84d3ab2a49406008c34e48a9420b8bd0e0ab4c5b9abd79dcdf822edb2b"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.545973 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.550019 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" event={"ID":"32ce1ea1-c409-4850-855b-fa8316d0d5e6","Type":"ContainerStarted","Data":"d6981a5564fe1d16e6e0c4052c9ed46cd3b13e7d4546a16202224887444021d2"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.550059 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" event={"ID":"32ce1ea1-c409-4850-855b-fa8316d0d5e6","Type":"ContainerStarted","Data":"34e66d36b50e80b9061f592a4decfd9c7d4f444bcc9ef0bd4c8f32e06eb73c1b"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.553062 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" event={"ID":"7d163c22-5dce-4cae-b536-c86dabcf6a91","Type":"ContainerStarted","Data":"248bcf3248873b2a33576586995abe7b58d84722213ee73c9d4e2c52ef7c79e7"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.554396 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.555365 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r9nf" event={"ID":"c207de79-b8c1-47f5-8e4b-5f320d760588","Type":"ContainerStarted","Data":"e9b7a8e27b131e68945eefb0580ac4e63c4c518d216dbdad96dc0ec1c1e7e625"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.558479 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" event={"ID":"54366321-671a-46e6-a4ca-a834ba087059","Type":"ContainerStarted","Data":"6aadee399530c6f0ec35666e11ff94e95c98500227fde737da37c3b05bd19c6c"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.558516 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" event={"ID":"54366321-671a-46e6-a4ca-a834ba087059","Type":"ContainerStarted","Data":"c27cff710f5afae282dbba5d684002ba3bf93a0ae98fec0d278aad2e372d15d6"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.559879 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" event={"ID":"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd","Type":"ContainerStarted","Data":"4312cf178fe9257626ecd668fe47a9fa1c0220103230fa54240398a6e19725f9"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.560864 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.561103 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.061072753 +0000 UTC m=+144.236808869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.561451 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.561910 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.061880676 +0000 UTC m=+144.237616792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.562234 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.569655 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q72r8" event={"ID":"e95bef62-5bd2-4a64-b1c5-df780f79ac51","Type":"ContainerStarted","Data":"1d940c3e97ceb6fa0ea09e9608e7e2cbe4bcf5abf8fe09213e0ef0e939a492d1"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.569706 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q72r8" event={"ID":"e95bef62-5bd2-4a64-b1c5-df780f79ac51","Type":"ContainerStarted","Data":"bb58b3aaa7374f7e480c4ecbb87863b8090e3ec0b39d0cbc018649a3c46e1499"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.570250 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.573061 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" event={"ID":"b28c51f3-e546-4bee-923f-e27b8c7c7567","Type":"ContainerStarted","Data":"28a76d5ba0a7cdfc8262d8b5902339bea8737bcd810a2d583f19b228b4e8ea44"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.573087 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" event={"ID":"b28c51f3-e546-4bee-923f-e27b8c7c7567","Type":"ContainerStarted","Data":"a397e5867bb8b519d22c348ed3179637f353c41418b15149e646fe141eac0806"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.579392 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.581282 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.585941 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" event={"ID":"f8056a9f-96dd-4c77-902d-0862fbc3b1ec","Type":"ContainerStarted","Data":"1dd6f15f8794faeec1f1483d491832fec6937f15bf3031b2f0fc451ec18e4dae"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.591630 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.594961 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" event={"ID":"a3ed3536-f707-47a5-b3df-3ec72bec773b","Type":"ContainerStarted","Data":"071fdaf3c50efd2a7976aab5f915496ca6540743d6a707be4cc7d0cb5c4ec926"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.595111 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" event={"ID":"a3ed3536-f707-47a5-b3df-3ec72bec773b","Type":"ContainerStarted","Data":"4ae891340b065f1d228b2317edd9199de9ff971494bb9ad1770418735e63c653"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.595272 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.597630 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" event={"ID":"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c","Type":"ContainerStarted","Data":"ad6a88b5458481337f8f8fee3e18c5f6945f7f535564caa59f2db015f0cd0abc"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.597666 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" event={"ID":"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c","Type":"ContainerStarted","Data":"a9949c4c1fcc8c61d39103a785a8ebd17877947e92ecfd1740377b683709f359"} Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.598230 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.600057 4702 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-n9lq9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.600122 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.600134 4702 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5dwsj container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.600222 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" podUID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.605278 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.631984 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9k26f"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.633006 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-465wh" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.662081 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.664659 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.164634095 +0000 UTC m=+144.340370211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.752464 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t56br"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.769156 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.773544 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.273247553 +0000 UTC m=+144.448983669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.797486 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6gdcn"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.802803 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.806729 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.818305 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.824166 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.824218 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.872618 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.873194 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.37315827 +0000 UTC m=+144.548894386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.873264 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.873724 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.373705896 +0000 UTC m=+144.549442012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.945850 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.958194 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mbn6w"] Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.974429 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.974925 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.47487408 +0000 UTC m=+144.650610196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:27 crc kubenswrapper[4702]: I1007 12:26:27.975058 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:27 crc kubenswrapper[4702]: E1007 12:26:27.975607 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.47558951 +0000 UTC m=+144.651325626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.018738 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.040736 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jk8m2"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.066380 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.077520 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.077867 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.577846285 +0000 UTC m=+144.753582401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.129918 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2"] Oct 07 12:26:28 crc kubenswrapper[4702]: W1007 12:26:28.133037 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5101fb34_4898_4acc_916d_3e2d69a46c1f.slice/crio-ba667d045cb3237edd47ee347c4fb3bbd01eebde2e79d6c548c3deb161e1d90a WatchSource:0}: Error finding container ba667d045cb3237edd47ee347c4fb3bbd01eebde2e79d6c548c3deb161e1d90a: Status 404 returned error can't find the container with id ba667d045cb3237edd47ee347c4fb3bbd01eebde2e79d6c548c3deb161e1d90a Oct 07 12:26:28 crc kubenswrapper[4702]: W1007 12:26:28.134031 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10b7a68d_a361_4fef_b4de_761c0cb9ed5b.slice/crio-fd4936d40ca7bdd76ed52b03bb486579278898c77c249c90a0a0351c83e2a95a WatchSource:0}: Error finding container fd4936d40ca7bdd76ed52b03bb486579278898c77c249c90a0a0351c83e2a95a: Status 404 returned error can't find the container with id fd4936d40ca7bdd76ed52b03bb486579278898c77c249c90a0a0351c83e2a95a Oct 07 12:26:28 crc kubenswrapper[4702]: W1007 12:26:28.148148 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod529d8a79_e1d3_4001_b791_4c440487bad9.slice/crio-88d9f67a9ec418691541f812236f7ac22abf39c51b07e8996d0ae27e35a76583 WatchSource:0}: Error finding container 88d9f67a9ec418691541f812236f7ac22abf39c51b07e8996d0ae27e35a76583: Status 404 returned error can't find the container with id 88d9f67a9ec418691541f812236f7ac22abf39c51b07e8996d0ae27e35a76583 Oct 07 12:26:28 crc kubenswrapper[4702]: W1007 12:26:28.149979 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56bcfcc7_401b_4261_b7de_2a28515c6267.slice/crio-4e685080fa28692863bed972de70054db74b124eb4dec1f2611370039409921f WatchSource:0}: Error finding container 4e685080fa28692863bed972de70054db74b124eb4dec1f2611370039409921f: Status 404 returned error can't find the container with id 4e685080fa28692863bed972de70054db74b124eb4dec1f2611370039409921f Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.183088 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.183641 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.683622571 +0000 UTC m=+144.859358687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.291752 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.292258 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.792236199 +0000 UTC m=+144.967972315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: W1007 12:26:28.306826 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5da410ed_29be_4a38_b486_89f0dd0252ae.slice/crio-0610ca27b55e929da01a90e1b0d9158a5a1e47e4a9997c428e731a5b98ec70b9 WatchSource:0}: Error finding container 0610ca27b55e929da01a90e1b0d9158a5a1e47e4a9997c428e731a5b98ec70b9: Status 404 returned error can't find the container with id 0610ca27b55e929da01a90e1b0d9158a5a1e47e4a9997c428e731a5b98ec70b9 Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.374716 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-q72r8" podStartSLOduration=119.374692794 podStartE2EDuration="1m59.374692794s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:28.373236042 +0000 UTC m=+144.548972178" watchObservedRunningTime="2025-10-07 12:26:28.374692794 +0000 UTC m=+144.550428920" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.394303 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.394768 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.894749102 +0000 UTC m=+145.070485218 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.495313 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.495720 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.995665548 +0000 UTC m=+145.171401664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.496014 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.496462 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:28.99644123 +0000 UTC m=+145.172177346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.499206 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qtvlc" podStartSLOduration=119.499186339 podStartE2EDuration="1m59.499186339s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:28.496365668 +0000 UTC m=+144.672101784" watchObservedRunningTime="2025-10-07 12:26:28.499186339 +0000 UTC m=+144.674922455" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.533150 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-k4phl"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.580836 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-78mw8" podStartSLOduration=119.58079857 podStartE2EDuration="1m59.58079857s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:28.57801397 +0000 UTC m=+144.753750086" watchObservedRunningTime="2025-10-07 12:26:28.58079857 +0000 UTC m=+144.756534676" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.590564 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.597042 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.597180 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.097149051 +0000 UTC m=+145.272885167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.598456 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.599321 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.099304303 +0000 UTC m=+145.275040419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.621976 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcw2q"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.631824 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t56br" event={"ID":"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b","Type":"ContainerStarted","Data":"4e395dc004a1c8f7ea09e3048d81e3df595bdfeb28506a14f2d9276560b4dcdb"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.636093 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.636890 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-b6wlp" podStartSLOduration=119.636880685 podStartE2EDuration="1m59.636880685s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:28.61415621 +0000 UTC m=+144.789892346" watchObservedRunningTime="2025-10-07 12:26:28.636880685 +0000 UTC m=+144.812616801" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.638377 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" event={"ID":"7758de03-0df5-45b4-8b7b-2a071740b2e0","Type":"ContainerStarted","Data":"b7c30f6b53e78fdcacd06f8db19c80903806e77c6f8331e7c3c8d1437a4c220c"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.656276 4702 generic.go:334] "Generic (PLEG): container finished" podID="e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd" containerID="2e0f8c550029d916396d524e48bd26591f5c14e4b644bcc7f565b2b35bf90bc8" exitCode=0 Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.667405 4702 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-n9lq9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.667455 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691362 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" event={"ID":"f8056a9f-96dd-4c77-902d-0862fbc3b1ec","Type":"ContainerStarted","Data":"84f1aedaefb3a3a703c84d7008608ec14e7f306808d69bdecdc8b442ccaf194e"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691461 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691484 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" event={"ID":"364be77d-ab70-49ff-9fa4-2b42f49b718b","Type":"ContainerStarted","Data":"b5e26f68a13c510b3ce746633acc427a168512fc93679a2edd036d5330e572b9"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691500 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" event={"ID":"5da410ed-29be-4a38-b486-89f0dd0252ae","Type":"ContainerStarted","Data":"0610ca27b55e929da01a90e1b0d9158a5a1e47e4a9997c428e731a5b98ec70b9"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691513 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" event={"ID":"a761b6de-3f12-4781-9a64-50bc79d6f482","Type":"ContainerStarted","Data":"dbbffacc5decf2faf898c0538f58cc7b5dce742e6ecb70e5fe1a4c0e751827c4"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691526 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" event={"ID":"56bcfcc7-401b-4261-b7de-2a28515c6267","Type":"ContainerStarted","Data":"4e685080fa28692863bed972de70054db74b124eb4dec1f2611370039409921f"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691538 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" event={"ID":"91bd4d74-eea0-42d9-9cb6-52c062a2b04c","Type":"ContainerStarted","Data":"f1d25d95fd73c23d7fc1671b228d427bf898b27ae5ec083640a76b04258bf4e2"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691550 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-465wh" event={"ID":"183df67a-7508-4b77-ac96-ec05258e68ee","Type":"ContainerStarted","Data":"0c8f9e5b1e671bb561959c4f6beff6706b63fe307b416b7c76931260570edf3f"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691563 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" event={"ID":"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a","Type":"ContainerStarted","Data":"077f1f1906b742b03063a4fc6440fa115077cdc2b8cdd97c92b41e4bdabaf8f8"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691574 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" event={"ID":"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd","Type":"ContainerDied","Data":"2e0f8c550029d916396d524e48bd26591f5c14e4b644bcc7f565b2b35bf90bc8"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691589 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" event={"ID":"e4cdf1f3-c551-4561-b731-a5413a13a724","Type":"ContainerStarted","Data":"64dca296d0af3ebfe3738cd5fb05ab104f61ea94b2c5b66ead74b70ebe0902a8"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691605 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" event={"ID":"529d8a79-e1d3-4001-b791-4c440487bad9","Type":"ContainerStarted","Data":"88d9f67a9ec418691541f812236f7ac22abf39c51b07e8996d0ae27e35a76583"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691619 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" event={"ID":"10b7a68d-a361-4fef-b4de-761c0cb9ed5b","Type":"ContainerStarted","Data":"fd4936d40ca7bdd76ed52b03bb486579278898c77c249c90a0a0351c83e2a95a"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691632 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jk8m2" event={"ID":"5101fb34-4898-4acc-916d-3e2d69a46c1f","Type":"ContainerStarted","Data":"ba667d045cb3237edd47ee347c4fb3bbd01eebde2e79d6c548c3deb161e1d90a"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691644 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" event={"ID":"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443","Type":"ContainerStarted","Data":"d3fa99babb883328f0640f3a82f43cdc9e37ed863608e6a3d2876e16f1f16c0e"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691657 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" event={"ID":"5ccb5e91-d8bc-4272-8656-d05e57e1d79c","Type":"ContainerStarted","Data":"b88750e6297d91da62b7b4010657cd038719c6c8aca72de5662e91bc47ef7c05"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691669 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" event={"ID":"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6","Type":"ContainerStarted","Data":"1fa81b0e6dda27016aa15169ee81d64d6124eee2f0c625c4be713ad7628e6305"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.691682 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n66bc" event={"ID":"3602b00e-3c66-4dd4-94fb-cc4421ea7566","Type":"ContainerStarted","Data":"4ce0bb68f48ea26f9abaa57b7d4406fe8e6b21fac56dfe1d050675a7c6ba45c0"} Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.705646 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.706198 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.20616777 +0000 UTC m=+145.381903886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.722104 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn"] Oct 07 12:26:28 crc kubenswrapper[4702]: W1007 12:26:28.757278 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod802177f9_2e2e_4cb8_895f_686ed017c9e3.slice/crio-a760d6de07df45c2c277691a1a58f5beb34c121c9539745d96afeb3e79660b5b WatchSource:0}: Error finding container a760d6de07df45c2c277691a1a58f5beb34c121c9539745d96afeb3e79660b5b: Status 404 returned error can't find the container with id a760d6de07df45c2c277691a1a58f5beb34c121c9539745d96afeb3e79660b5b Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.807405 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.812844 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.312822832 +0000 UTC m=+145.488558948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.896506 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c5n5d"] Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.912618 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:28 crc kubenswrapper[4702]: E1007 12:26:28.913178 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.413142921 +0000 UTC m=+145.588879037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.933134 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j8v4k" podStartSLOduration=119.933114916 podStartE2EDuration="1m59.933114916s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:28.932608022 +0000 UTC m=+145.108344138" watchObservedRunningTime="2025-10-07 12:26:28.933114916 +0000 UTC m=+145.108851032" Oct 07 12:26:28 crc kubenswrapper[4702]: I1007 12:26:28.977265 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-w7whn"] Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.017226 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.038224 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.538184772 +0000 UTC m=+145.713920888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.102997 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92"] Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.111541 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" podStartSLOduration=120.111521744 podStartE2EDuration="2m0.111521744s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:29.109350771 +0000 UTC m=+145.285086887" watchObservedRunningTime="2025-10-07 12:26:29.111521744 +0000 UTC m=+145.287257860" Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.118768 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24"] Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.119609 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.120262 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.620238665 +0000 UTC m=+145.795974781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.148924 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl"] Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.149015 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" podStartSLOduration=120.148988613 podStartE2EDuration="2m0.148988613s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:29.13638017 +0000 UTC m=+145.312116286" watchObservedRunningTime="2025-10-07 12:26:29.148988613 +0000 UTC m=+145.324724729" Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.208731 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5bbhq" podStartSLOduration=120.208695472 podStartE2EDuration="2m0.208695472s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:29.195116091 +0000 UTC m=+145.370852237" watchObservedRunningTime="2025-10-07 12:26:29.208695472 +0000 UTC m=+145.384431588" Oct 07 12:26:29 crc kubenswrapper[4702]: W1007 12:26:29.221352 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9596f202_95f9_45ae_92be_6162cc42a574.slice/crio-424cf9b3d2f824b55906c55139a29e136d3268e275d007a240f27a0685a2f947 WatchSource:0}: Error finding container 424cf9b3d2f824b55906c55139a29e136d3268e275d007a240f27a0685a2f947: Status 404 returned error can't find the container with id 424cf9b3d2f824b55906c55139a29e136d3268e275d007a240f27a0685a2f947 Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.222751 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.224426 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.724399535 +0000 UTC m=+145.900135651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: W1007 12:26:29.256444 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf6cffa0_4366_4806_a956_34437988bdcf.slice/crio-baf38884d04302e25fa740022934bbe6aae2e312c32787bdd3752b0bbd8f8934 WatchSource:0}: Error finding container baf38884d04302e25fa740022934bbe6aae2e312c32787bdd3752b0bbd8f8934: Status 404 returned error can't find the container with id baf38884d04302e25fa740022934bbe6aae2e312c32787bdd3752b0bbd8f8934 Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.302709 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hwfzn" podStartSLOduration=120.302682329 podStartE2EDuration="2m0.302682329s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:29.301745132 +0000 UTC m=+145.477481268" watchObservedRunningTime="2025-10-07 12:26:29.302682329 +0000 UTC m=+145.478418445" Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.325876 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.326532 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.826509775 +0000 UTC m=+146.002245891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.350061 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt"] Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.428152 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.428665 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:29.928650447 +0000 UTC m=+146.104386563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.528882 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.529228 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.029202623 +0000 UTC m=+146.204938739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.530109 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.030092988 +0000 UTC m=+146.205829104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.530696 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:29 crc kubenswrapper[4702]: W1007 12:26:29.534238 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f7df49_3241_4dd4_b579_ac0d7436f9ab.slice/crio-b28ccfd05268242ac8c69ad0caa7547b87850851f9c9fe30aeabdacb6728889d WatchSource:0}: Error finding container b28ccfd05268242ac8c69ad0caa7547b87850851f9c9fe30aeabdacb6728889d: Status 404 returned error can't find the container with id b28ccfd05268242ac8c69ad0caa7547b87850851f9c9fe30aeabdacb6728889d Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.632799 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.633211 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.133188637 +0000 UTC m=+146.308924753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.633498 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.633825 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.133812535 +0000 UTC m=+146.309548651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.701367 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jk8m2" event={"ID":"5101fb34-4898-4acc-916d-3e2d69a46c1f","Type":"ContainerStarted","Data":"78825ff29e27efd64bec129acbd28fd2bda901598623eee76c54aefe6e3ec512"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.709506 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" event={"ID":"7758de03-0df5-45b4-8b7b-2a071740b2e0","Type":"ContainerStarted","Data":"cbf729b48578eb80f15f8f6ebcca864244a6f73ba9b278f3a64927fb3a11d515"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.712590 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" event={"ID":"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06","Type":"ContainerStarted","Data":"6e631b8c392a80bd70406c5fc4eeb095124cb7bf768413c05610172a7492df76"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.714182 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" event={"ID":"d1c7c81c-7284-4c9c-96de-d7c546a8eb97","Type":"ContainerStarted","Data":"67fd74823365a974418650b499d9aff0e8f716907a6ee861b197530629ff4bc4"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.716320 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" event={"ID":"802177f9-2e2e-4cb8-895f-686ed017c9e3","Type":"ContainerStarted","Data":"a760d6de07df45c2c277691a1a58f5beb34c121c9539745d96afeb3e79660b5b"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.739398 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.739805 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.239773057 +0000 UTC m=+146.415509173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.740005 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-465wh" event={"ID":"183df67a-7508-4b77-ac96-ec05258e68ee","Type":"ContainerStarted","Data":"2643eb362e2514f033630a5a12191083cc26984e9c70b90bed2c1adaa81cf543"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.747801 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" event={"ID":"d8dd7d8a-f682-4492-8e4d-7318d345a9dd","Type":"ContainerStarted","Data":"afb8468ba449d751535d88f4d3d5356085eb34212781fb27e71cbc8dd8d0b5bc"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.763019 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" event={"ID":"6f164dd1-c728-4d35-b7f8-12fca172497d","Type":"ContainerStarted","Data":"e86f9002f87e74c8143589a5755ba1b1c8e510815c0be2c5859a5040d2912095"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.790847 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" event={"ID":"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a","Type":"ContainerStarted","Data":"627677457b2b091c0e38eda67bd0eb58479bd0d0a7e88ef8777acde732ac9db9"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.797803 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c5n5d" event={"ID":"c336b672-f2b9-45ce-94ea-e56ed02c3b9f","Type":"ContainerStarted","Data":"5adc044d51388d6c57f47278c67eaba27d8598f8263b2025dd1460412e9ea533"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.818987 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" event={"ID":"bf6cffa0-4366-4806-a956-34437988bdcf","Type":"ContainerStarted","Data":"baf38884d04302e25fa740022934bbe6aae2e312c32787bdd3752b0bbd8f8934"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.840946 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.842381 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.342348901 +0000 UTC m=+146.518085197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.857760 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" event={"ID":"e4cdf1f3-c551-4561-b731-a5413a13a724","Type":"ContainerStarted","Data":"589cdcdcce85417838b088c78f4707ac8ff3b4deb11f46b92ec2b725b1c5f010"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.857831 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" event={"ID":"e4cdf1f3-c551-4561-b731-a5413a13a724","Type":"ContainerStarted","Data":"daeedc09873160f42edeadfb06a79e1ac9a89464a341d3114dbc8e713174e650"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.900693 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" event={"ID":"a761b6de-3f12-4781-9a64-50bc79d6f482","Type":"ContainerStarted","Data":"3560a4bfb79d072843f72bcc450ae40b3f699dbfceb1f74f21ad230bb282a3ae"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.913198 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" event={"ID":"9596f202-95f9-45ae-92be-6162cc42a574","Type":"ContainerStarted","Data":"424cf9b3d2f824b55906c55139a29e136d3268e275d007a240f27a0685a2f947"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.932838 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n66bc" event={"ID":"3602b00e-3c66-4dd4-94fb-cc4421ea7566","Type":"ContainerStarted","Data":"9c359eeb07ad27ee99af0ec2e31cbdd72be01d7c3a98c230b785ee876181c6b5"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.949881 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" event={"ID":"9ca3c868-88b7-49e3-acd0-ba172b13b24f","Type":"ContainerStarted","Data":"6660dc8f2def02e8fa8f267c4f6c5679a72bd25f24a529d57afec1d770520ff3"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.979163 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:29 crc kubenswrapper[4702]: E1007 12:26:29.980158 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.480141289 +0000 UTC m=+146.655877405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.991465 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" event={"ID":"2a92879d-06fd-4c7f-853c-360e5215a574","Type":"ContainerStarted","Data":"15e005d690e3b19abb9609347ae2ca32949b5b4584e1e36b6ca37fd71ae4ebef"} Oct 07 12:26:29 crc kubenswrapper[4702]: I1007 12:26:29.992532 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" event={"ID":"48f7df49-3241-4dd4-b579-ac0d7436f9ab","Type":"ContainerStarted","Data":"b28ccfd05268242ac8c69ad0caa7547b87850851f9c9fe30aeabdacb6728889d"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:29.998775 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" event={"ID":"5ccb5e91-d8bc-4272-8656-d05e57e1d79c","Type":"ContainerStarted","Data":"35dea1f4e1d8856b33ef678c9f4930f90ab8769263404e1ee9eed38c93dca3ee"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.022023 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" event={"ID":"7d163c22-5dce-4cae-b536-c86dabcf6a91","Type":"ContainerStarted","Data":"40bd84bc3e5159be4d679b0fbf2e5c2993ec325b1f34c3c745430d03cc7f0430"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.045233 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" event={"ID":"364be77d-ab70-49ff-9fa4-2b42f49b718b","Type":"ContainerStarted","Data":"f390836ac8a5aa23517284250ef14965f14ad56bf4e266fb32fb752154359275"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.046126 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.063207 4702 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-lbqfj container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.063277 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" podUID="364be77d-ab70-49ff-9fa4-2b42f49b718b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.063844 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" event={"ID":"5da410ed-29be-4a38-b486-89f0dd0252ae","Type":"ContainerStarted","Data":"3b8e8a154d37c56f62c12ac2d608d38e754d1d2761135199c5bf625b9ddaeb7c"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.071671 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r9nf" event={"ID":"c207de79-b8c1-47f5-8e4b-5f320d760588","Type":"ContainerStarted","Data":"aa2d2154d425081d0baf8882f174459d01e54293b808868014acab3727ee2b7f"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.076411 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.086017 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-465wh" podStartSLOduration=6.085993517 podStartE2EDuration="6.085993517s" podCreationTimestamp="2025-10-07 12:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.062945884 +0000 UTC m=+146.238682000" watchObservedRunningTime="2025-10-07 12:26:30.085993517 +0000 UTC m=+146.261729633" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.086649 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2mpvg" podStartSLOduration=121.086645416 podStartE2EDuration="2m1.086645416s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.084024761 +0000 UTC m=+146.259760877" watchObservedRunningTime="2025-10-07 12:26:30.086645416 +0000 UTC m=+146.262381532" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.087543 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.089680 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.589656493 +0000 UTC m=+146.765392609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.092196 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r9nf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.092262 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r9nf" podUID="c207de79-b8c1-47f5-8e4b-5f320d760588" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.114988 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t56br" event={"ID":"dc1f4290-b2cb-469b-9fa2-94f7ab35f42b","Type":"ContainerStarted","Data":"9eb718b7b5f7fa7543275c57783642c4f3e57c87e82615cfe969b3b8adb416f1"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.117722 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.133139 4702 patch_prober.go:28] interesting pod/console-operator-58897d9998-t56br container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.133232 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t56br" podUID="dc1f4290-b2cb-469b-9fa2-94f7ab35f42b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.136752 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" event={"ID":"529d8a79-e1d3-4001-b791-4c440487bad9","Type":"ContainerStarted","Data":"189a779c6ba6424773b536326fa0cc0aa84a8d70baa2abbead7b41f34207e9e7"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.144707 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" event={"ID":"8c414497-e825-484c-bfb1-1e9dc5e4d9d2","Type":"ContainerStarted","Data":"5d912b59b1d3b713ba5a58119885ee2859f58083115b4afa7be68ec0074c8dbf"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.163929 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-n66bc" podStartSLOduration=121.163889391 podStartE2EDuration="2m1.163889391s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.162936003 +0000 UTC m=+146.338672119" watchObservedRunningTime="2025-10-07 12:26:30.163889391 +0000 UTC m=+146.339625497" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.168872 4702 generic.go:334] "Generic (PLEG): container finished" podID="b7763d6d-7a4e-4aa3-80c0-fc4e27a66443" containerID="cd68c32cf5447f8773cc15cc6b26684beb0167bc5936a1097f47318407376e1e" exitCode=0 Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.172022 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" event={"ID":"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443","Type":"ContainerDied","Data":"cd68c32cf5447f8773cc15cc6b26684beb0167bc5936a1097f47318407376e1e"} Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.191630 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.192157 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.692116074 +0000 UTC m=+146.867852190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.289752 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n5p5n" podStartSLOduration=121.289728954 podStartE2EDuration="2m1.289728954s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.248191798 +0000 UTC m=+146.423927914" watchObservedRunningTime="2025-10-07 12:26:30.289728954 +0000 UTC m=+146.465465070" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.290593 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jk8m2" podStartSLOduration=6.290586598 podStartE2EDuration="6.290586598s" podCreationTimestamp="2025-10-07 12:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.289400934 +0000 UTC m=+146.465137050" watchObservedRunningTime="2025-10-07 12:26:30.290586598 +0000 UTC m=+146.466322714" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.299824 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.306815 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.806799095 +0000 UTC m=+146.982535211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.341569 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-t56br" podStartSLOduration=121.341549466 podStartE2EDuration="2m1.341549466s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.339781605 +0000 UTC m=+146.515517721" watchObservedRunningTime="2025-10-07 12:26:30.341549466 +0000 UTC m=+146.517285582" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.402873 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.403873 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:30.90385674 +0000 UTC m=+147.079592856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.428616 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" podStartSLOduration=121.428602223 podStartE2EDuration="2m1.428602223s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.427065019 +0000 UTC m=+146.602801125" watchObservedRunningTime="2025-10-07 12:26:30.428602223 +0000 UTC m=+146.604338339" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.461782 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-q5qbc" podStartSLOduration=121.461756178 podStartE2EDuration="2m1.461756178s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.460812101 +0000 UTC m=+146.636548227" watchObservedRunningTime="2025-10-07 12:26:30.461756178 +0000 UTC m=+146.637492294" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.464714 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.473111 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:30 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:30 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:30 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.473175 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.500982 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9r9nf" podStartSLOduration=121.500959897 podStartE2EDuration="2m1.500959897s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.500428082 +0000 UTC m=+146.676164198" watchObservedRunningTime="2025-10-07 12:26:30.500959897 +0000 UTC m=+146.676696013" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.508149 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.508578 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.008560046 +0000 UTC m=+147.184296162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.537289 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-t72fg" podStartSLOduration=121.537268263 podStartE2EDuration="2m1.537268263s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.53369782 +0000 UTC m=+146.709433936" watchObservedRunningTime="2025-10-07 12:26:30.537268263 +0000 UTC m=+146.713004379" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.580153 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5x5r2" podStartSLOduration=121.580127287 podStartE2EDuration="2m1.580127287s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.570326595 +0000 UTC m=+146.746062711" watchObservedRunningTime="2025-10-07 12:26:30.580127287 +0000 UTC m=+146.755863403" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.610835 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.611334 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.111307775 +0000 UTC m=+147.287043891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.702470 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" podStartSLOduration=121.70244718 podStartE2EDuration="2m1.70244718s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:30.700425931 +0000 UTC m=+146.876162047" watchObservedRunningTime="2025-10-07 12:26:30.70244718 +0000 UTC m=+146.878183296" Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.722666 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.723195 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.223177616 +0000 UTC m=+147.398913722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.824856 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.825599 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.325556985 +0000 UTC m=+147.501293101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.825818 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.826192 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.326176653 +0000 UTC m=+147.501912769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:30 crc kubenswrapper[4702]: I1007 12:26:30.931615 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:30 crc kubenswrapper[4702]: E1007 12:26:30.932006 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.43197557 +0000 UTC m=+147.607711686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.032935 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.033511 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.533491933 +0000 UTC m=+147.709228049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.138519 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.139161 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.639132956 +0000 UTC m=+147.814869082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.189862 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" event={"ID":"6f164dd1-c728-4d35-b7f8-12fca172497d","Type":"ContainerStarted","Data":"e384fa09c91796f1e92901132c8b696d6d9e0947e072110f88ac54ffbf2a36e9"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.190866 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.199702 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" event={"ID":"cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06","Type":"ContainerStarted","Data":"1651accdb23ae51d17a7fafe4bb69913ae5901c85c6dab79c017087ba8145d69"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.201344 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.212124 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" event={"ID":"802177f9-2e2e-4cb8-895f-686ed017c9e3","Type":"ContainerStarted","Data":"07f9ea168e0bfac4ac9aea619a1dd96fc8cf7a913b6b1f1c78496659d8b4666d"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.214341 4702 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pcw2q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.214392 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.224055 4702 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2drjl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.224146 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" podUID="cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.224566 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" event={"ID":"2eb394c2-b11a-4466-a9aa-708bbb4ac4e6","Type":"ContainerStarted","Data":"f5447db15850e09f23fefea15716bedc73c0216dd8762f72de5f8e522a87f220"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.236770 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" podStartSLOduration=122.236745677 podStartE2EDuration="2m2.236745677s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.236414057 +0000 UTC m=+147.412150173" watchObservedRunningTime="2025-10-07 12:26:31.236745677 +0000 UTC m=+147.412481793" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.240434 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.243664 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.743646545 +0000 UTC m=+147.919382661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.269167 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" event={"ID":"9ca3c868-88b7-49e3-acd0-ba172b13b24f","Type":"ContainerStarted","Data":"654ae82df2c77249b30e84f7450c2809a746cd5d80b4978f0efa6cf495fc20ae"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.277555 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7kg" podStartSLOduration=122.277532441 podStartE2EDuration="2m2.277532441s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.275594875 +0000 UTC m=+147.451330981" watchObservedRunningTime="2025-10-07 12:26:31.277532441 +0000 UTC m=+147.453268557" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.281581 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c5n5d" event={"ID":"c336b672-f2b9-45ce-94ea-e56ed02c3b9f","Type":"ContainerStarted","Data":"8e01c1f6588890e0ced6d34793000be29be02c1c99fa140cdcb87f3f04aa8650"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.287959 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" event={"ID":"bf6cffa0-4366-4806-a956-34437988bdcf","Type":"ContainerStarted","Data":"ba7907d1a66bd2911e3e45e6dc678382d16cf44586c79a8e3a15f14fdbd65990"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.340926 4702 generic.go:334] "Generic (PLEG): container finished" podID="91bd4d74-eea0-42d9-9cb6-52c062a2b04c" containerID="0eeec687010af0356a92a360eaff930ca9ae842d760e8d53dbd789818f5595cb" exitCode=0 Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.341053 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" event={"ID":"91bd4d74-eea0-42d9-9cb6-52c062a2b04c","Type":"ContainerDied","Data":"0eeec687010af0356a92a360eaff930ca9ae842d760e8d53dbd789818f5595cb"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.349096 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.350367 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.850344738 +0000 UTC m=+148.026080854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.384413 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" event={"ID":"48f7df49-3241-4dd4-b579-ac0d7436f9ab","Type":"ContainerStarted","Data":"4b0238231d93ee2098c4bd172735c8722c571aaacd540416ec0c8c161262b0d1"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.385298 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.405048 4702 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-b8hkt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.405117 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" podUID="48f7df49-3241-4dd4-b579-ac0d7436f9ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.410481 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" event={"ID":"9596f202-95f9-45ae-92be-6162cc42a574","Type":"ContainerStarted","Data":"299d554a24eb209b437f385bb41fa2a40778feff4c29128b96002267735a9d99"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.411942 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" event={"ID":"b7763d6d-7a4e-4aa3-80c0-fc4e27a66443","Type":"ContainerStarted","Data":"87907d0729afeb31a9d273659718e54bea8e0bdcfc4eb183bad0aaccc4cc86cc"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.412631 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.429528 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" event={"ID":"2a92879d-06fd-4c7f-853c-360e5215a574","Type":"ContainerStarted","Data":"15189461e24f29424af4ee5e4586d21b9738f00ae60dafbcfc8dcf015c3cad80"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.430532 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.443090 4702 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cgb8d container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.443157 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" podUID="2a92879d-06fd-4c7f-853c-360e5215a574" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.451674 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.452166 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:31.95215257 +0000 UTC m=+148.127888686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.464491 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" event={"ID":"56bcfcc7-401b-4261-b7de-2a28515c6267","Type":"ContainerStarted","Data":"f0dc496be96d432edb6b4299ae80131ff898139fd05d5fdbfa642f1bbd8f8d73"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.464557 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" event={"ID":"56bcfcc7-401b-4261-b7de-2a28515c6267","Type":"ContainerStarted","Data":"83c1a54d950a0d4494e1412716c1fc8584065b742396dbc1e27e2f2ee2263823"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.477206 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:31 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:31 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:31 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.477293 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.479196 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" podStartSLOduration=122.479169758 podStartE2EDuration="2m2.479169758s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.368699307 +0000 UTC m=+147.544435413" watchObservedRunningTime="2025-10-07 12:26:31.479169758 +0000 UTC m=+147.654905864" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.483283 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" event={"ID":"a761b6de-3f12-4781-9a64-50bc79d6f482","Type":"ContainerStarted","Data":"f55914d6ac411a6f6b02627052da043ac796525fc2f17787c333db0fa0c26fec"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.513271 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" event={"ID":"2f4b3ec1-1351-41b7-a39c-27b29d6f8f2a","Type":"ContainerStarted","Data":"57730d615717268b92a9f0496af19daf130cd93dbc43a59d5be81e616471be33"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.527003 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-k4phl" podStartSLOduration=122.526987365 podStartE2EDuration="2m2.526987365s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.5264523 +0000 UTC m=+147.702188416" watchObservedRunningTime="2025-10-07 12:26:31.526987365 +0000 UTC m=+147.702723481" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.535373 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-47q4n" podStartSLOduration=122.535345526 podStartE2EDuration="2m2.535345526s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.492602385 +0000 UTC m=+147.668338501" watchObservedRunningTime="2025-10-07 12:26:31.535345526 +0000 UTC m=+147.711081642" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.536097 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" event={"ID":"10b7a68d-a361-4fef-b4de-761c0cb9ed5b","Type":"ContainerStarted","Data":"1be4dd664976b85b61ffb799c69fed8c829753bce2a4a8f3b1323c04195bf1ef"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.556598 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.558093 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.05806797 +0000 UTC m=+148.233804126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.560082 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" event={"ID":"8c414497-e825-484c-bfb1-1e9dc5e4d9d2","Type":"ContainerStarted","Data":"eb80ec7e5d45c30797cedf8ab73b65665a327dd2eaa5bf909e44dd83eea919ae"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.560138 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" event={"ID":"8c414497-e825-484c-bfb1-1e9dc5e4d9d2","Type":"ContainerStarted","Data":"619fd9c43068eb46d184b09161a68b5a1a183aa2a3127bdb8d70a7d96ca665db"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.596389 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r9nf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.596453 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r9nf" podUID="c207de79-b8c1-47f5-8e4b-5f320d760588" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.596518 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" event={"ID":"d1c7c81c-7284-4c9c-96de-d7c546a8eb97","Type":"ContainerStarted","Data":"9bc89aae50c6f292f73147797ffe269cd277ede701c0b3b7eb77655e3c55fe08"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.596563 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" event={"ID":"d1c7c81c-7284-4c9c-96de-d7c546a8eb97","Type":"ContainerStarted","Data":"e3ed4e5427499bbb41f43db6e9c24b9b30bc67a8e9666dabe2cafcd9a376ff5b"} Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.648697 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" podStartSLOduration=122.64867683 podStartE2EDuration="2m2.64867683s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.583785641 +0000 UTC m=+147.759521757" watchObservedRunningTime="2025-10-07 12:26:31.64867683 +0000 UTC m=+147.824412946" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.665868 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.669827 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.169806178 +0000 UTC m=+148.345542284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.710528 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" podStartSLOduration=122.71050694 podStartE2EDuration="2m2.71050694s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.650558094 +0000 UTC m=+147.826294210" watchObservedRunningTime="2025-10-07 12:26:31.71050694 +0000 UTC m=+147.886243056" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.767455 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.768079 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.268052968 +0000 UTC m=+148.443789084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.788124 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7lknr" podStartSLOduration=122.788102785 podStartE2EDuration="2m2.788102785s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.78551219 +0000 UTC m=+147.961248306" watchObservedRunningTime="2025-10-07 12:26:31.788102785 +0000 UTC m=+147.963838901" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.788663 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-mbn6w" podStartSLOduration=122.788658101 podStartE2EDuration="2m2.788658101s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.706120754 +0000 UTC m=+147.881856860" watchObservedRunningTime="2025-10-07 12:26:31.788658101 +0000 UTC m=+147.964394217" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.820279 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" podStartSLOduration=122.820246641 podStartE2EDuration="2m2.820246641s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.81988735 +0000 UTC m=+147.995623466" watchObservedRunningTime="2025-10-07 12:26:31.820246641 +0000 UTC m=+147.995982757" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.875328 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.875628 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.375591385 +0000 UTC m=+148.551327501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.926825 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szgpc" podStartSLOduration=122.926801719 podStartE2EDuration="2m2.926801719s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.925505272 +0000 UTC m=+148.101241388" watchObservedRunningTime="2025-10-07 12:26:31.926801719 +0000 UTC m=+148.102537825" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.928857 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6gdcn" podStartSLOduration=122.928834178 podStartE2EDuration="2m2.928834178s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.872528816 +0000 UTC m=+148.048264932" watchObservedRunningTime="2025-10-07 12:26:31.928834178 +0000 UTC m=+148.104570294" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.978993 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dcpzn" podStartSLOduration=122.978971632 podStartE2EDuration="2m2.978971632s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:31.978411316 +0000 UTC m=+148.154147432" watchObservedRunningTime="2025-10-07 12:26:31.978971632 +0000 UTC m=+148.154707748" Oct 07 12:26:31 crc kubenswrapper[4702]: I1007 12:26:31.979435 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:31 crc kubenswrapper[4702]: E1007 12:26:31.979946 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.479923059 +0000 UTC m=+148.655659175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.036679 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z4722" podStartSLOduration=123.036662213 podStartE2EDuration="2m3.036662213s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:32.031794053 +0000 UTC m=+148.207530169" watchObservedRunningTime="2025-10-07 12:26:32.036662213 +0000 UTC m=+148.212398329" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.081980 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.082748 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.582361549 +0000 UTC m=+148.758097665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.183795 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.184309 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.684282624 +0000 UTC m=+148.860018740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.251456 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.285279 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.285788 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.785750067 +0000 UTC m=+148.961486183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.386361 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.387191 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.887161187 +0000 UTC m=+149.062897303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.475634 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:32 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:32 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:32 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.475772 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.488167 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.488553 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:32.988537947 +0000 UTC m=+149.164274063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.589261 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.589559 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.089486534 +0000 UTC m=+149.265222650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.589969 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.590463 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.090442751 +0000 UTC m=+149.266178867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.597184 4702 patch_prober.go:28] interesting pod/console-operator-58897d9998-t56br container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.597278 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t56br" podUID="dc1f4290-b2cb-469b-9fa2-94f7ab35f42b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.604048 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" event={"ID":"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd","Type":"ContainerStarted","Data":"7ec9057e1eeff971281c0c2deb8d12e1340ebaae71a3c38db87151b6be38e536"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.604114 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" event={"ID":"e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd","Type":"ContainerStarted","Data":"6bf245cde257fdd0811f9fa610af7a4a3e3b006a85609fd0dda86d766a2d5371"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.608005 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c5n5d" event={"ID":"c336b672-f2b9-45ce-94ea-e56ed02c3b9f","Type":"ContainerStarted","Data":"25890e77ff735e0e9bed2001e83a5be0312d9106324ea3e09a3c4fc1b16a6446"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.609127 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.612011 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" event={"ID":"bf6cffa0-4366-4806-a956-34437988bdcf","Type":"ContainerStarted","Data":"afdb60ae709e4a2ad34e9948f6fb1cbeec651e19f9099c59a55ca05d1b5f1e24"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.612721 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.622183 4702 generic.go:334] "Generic (PLEG): container finished" podID="5da410ed-29be-4a38-b486-89f0dd0252ae" containerID="3b8e8a154d37c56f62c12ac2d608d38e754d1d2761135199c5bf625b9ddaeb7c" exitCode=0 Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.622373 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" event={"ID":"5da410ed-29be-4a38-b486-89f0dd0252ae","Type":"ContainerDied","Data":"3b8e8a154d37c56f62c12ac2d608d38e754d1d2761135199c5bf625b9ddaeb7c"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.624472 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" event={"ID":"d8dd7d8a-f682-4492-8e4d-7318d345a9dd","Type":"ContainerStarted","Data":"0adbff60f56cd4d36f2ddebedc1b4da7de0dcdea09b6f3573ecc4588fb17c58f"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.627218 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" event={"ID":"91bd4d74-eea0-42d9-9cb6-52c062a2b04c","Type":"ContainerStarted","Data":"bde8fbf2e4f8e2691673131ceefc39a192038c43ebf7ec79b3b2caf3b0be8385"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.629557 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" event={"ID":"9596f202-95f9-45ae-92be-6162cc42a574","Type":"ContainerStarted","Data":"84436634fcad17b183c6a0d6a4b69468428750ef6266bb61f4f318240348733f"} Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.630712 4702 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-b8hkt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.630782 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" podUID="48f7df49-3241-4dd4-b579-ac0d7436f9ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.631215 4702 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pcw2q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.631275 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.631419 4702 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cgb8d container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.631472 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" podUID="2a92879d-06fd-4c7f-853c-360e5215a574" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.631613 4702 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2drjl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.631648 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" podUID="cd7e8c79-c0ed-4e7e-ac14-e0b3b5ac9f06" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.686408 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" podStartSLOduration=123.686386764 podStartE2EDuration="2m3.686386764s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:32.668383636 +0000 UTC m=+148.844119752" watchObservedRunningTime="2025-10-07 12:26:32.686386764 +0000 UTC m=+148.862122880" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.691951 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.692199 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.692672 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.692820 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.693115 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.706211 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.707796 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.711087 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.713322 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.213294229 +0000 UTC m=+149.389030525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.716196 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.788943 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gnc24" podStartSLOduration=123.788888946 podStartE2EDuration="2m3.788888946s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:32.756098212 +0000 UTC m=+148.931834338" watchObservedRunningTime="2025-10-07 12:26:32.788888946 +0000 UTC m=+148.964625062" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.795541 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.795988 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.29596691 +0000 UTC m=+149.471703026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.865947 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" podStartSLOduration=123.865923935 podStartE2EDuration="2m3.865923935s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:32.863817874 +0000 UTC m=+149.039553990" watchObservedRunningTime="2025-10-07 12:26:32.865923935 +0000 UTC m=+149.041660051" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.897232 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.897459 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.397423682 +0000 UTC m=+149.573159798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.897737 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:32 crc kubenswrapper[4702]: E1007 12:26:32.898132 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.398117382 +0000 UTC m=+149.573853498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.973403 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.986512 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 12:26:32 crc kubenswrapper[4702]: I1007 12:26:32.997631 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.000555 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.000741 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.500696186 +0000 UTC m=+149.676432312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.001033 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.001472 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.501454658 +0000 UTC m=+149.677190774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.018356 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" podStartSLOduration=124.018322444 podStartE2EDuration="2m4.018322444s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:32.963187526 +0000 UTC m=+149.138923652" watchObservedRunningTime="2025-10-07 12:26:33.018322444 +0000 UTC m=+149.194058560" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.065796 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-c5n5d" podStartSLOduration=9.065766 podStartE2EDuration="9.065766s" podCreationTimestamp="2025-10-07 12:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:33.021819214 +0000 UTC m=+149.197555330" watchObservedRunningTime="2025-10-07 12:26:33.065766 +0000 UTC m=+149.241502116" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.102151 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.102395 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.602356834 +0000 UTC m=+149.778092950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.102460 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.103032 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.603008202 +0000 UTC m=+149.778744318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.207462 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.208224 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.708200272 +0000 UTC m=+149.883936388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.311956 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.312374 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.812357691 +0000 UTC m=+149.988093807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.419227 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.440591 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.441192 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:33.941171251 +0000 UTC m=+150.116907367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.473359 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:33 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:33 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:33 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.473447 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.542636 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.543729 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.043710564 +0000 UTC m=+150.219446680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.643401 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.643722 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.143705984 +0000 UTC m=+150.319442100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.682549 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"321ba5aeed58e6c1cb2a74c580a56290b2fd9d0eb113b59d495dfbb06b36259e"} Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.683937 4702 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pcw2q container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.683999 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.745036 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.746980 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.246967338 +0000 UTC m=+150.422703454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.762987 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cgb8d" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.765270 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b8hkt" Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.846756 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.847158 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.347135691 +0000 UTC m=+150.522871807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:33 crc kubenswrapper[4702]: I1007 12:26:33.954554 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:33 crc kubenswrapper[4702]: E1007 12:26:33.955216 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.455204174 +0000 UTC m=+150.630940280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.061192 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.061538 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.561520255 +0000 UTC m=+150.737256361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.163461 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.163910 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.663880403 +0000 UTC m=+150.839616519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.264160 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.264270 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.764241093 +0000 UTC m=+150.939977209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.264442 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.264845 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.764836801 +0000 UTC m=+150.940572917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.268763 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.366406 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt47q\" (UniqueName: \"kubernetes.io/projected/5da410ed-29be-4a38-b486-89f0dd0252ae-kube-api-access-rt47q\") pod \"5da410ed-29be-4a38-b486-89f0dd0252ae\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.366459 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5da410ed-29be-4a38-b486-89f0dd0252ae-secret-volume\") pod \"5da410ed-29be-4a38-b486-89f0dd0252ae\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.366655 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.366724 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5da410ed-29be-4a38-b486-89f0dd0252ae-config-volume\") pod \"5da410ed-29be-4a38-b486-89f0dd0252ae\" (UID: \"5da410ed-29be-4a38-b486-89f0dd0252ae\") " Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.367357 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da410ed-29be-4a38-b486-89f0dd0252ae-config-volume" (OuterVolumeSpecName: "config-volume") pod "5da410ed-29be-4a38-b486-89f0dd0252ae" (UID: "5da410ed-29be-4a38-b486-89f0dd0252ae"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.370986 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.870933926 +0000 UTC m=+151.046670042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.378996 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da410ed-29be-4a38-b486-89f0dd0252ae-kube-api-access-rt47q" (OuterVolumeSpecName: "kube-api-access-rt47q") pod "5da410ed-29be-4a38-b486-89f0dd0252ae" (UID: "5da410ed-29be-4a38-b486-89f0dd0252ae"). InnerVolumeSpecName "kube-api-access-rt47q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.404253 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da410ed-29be-4a38-b486-89f0dd0252ae-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5da410ed-29be-4a38-b486-89f0dd0252ae" (UID: "5da410ed-29be-4a38-b486-89f0dd0252ae"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.468722 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.469186 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:34.969164495 +0000 UTC m=+151.144900611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.469303 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5da410ed-29be-4a38-b486-89f0dd0252ae-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.469318 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt47q\" (UniqueName: \"kubernetes.io/projected/5da410ed-29be-4a38-b486-89f0dd0252ae-kube-api-access-rt47q\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.469329 4702 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5da410ed-29be-4a38-b486-89f0dd0252ae-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.480265 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:34 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:34 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:34 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.480354 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.529882 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2drjl" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.571561 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.571916 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.071844992 +0000 UTC m=+151.247581108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.674720 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.675121 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.175103066 +0000 UTC m=+151.350839182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.684380 4702 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-9k26f container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.684444 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" podUID="b7763d6d-7a4e-4aa3-80c0-fc4e27a66443" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.751471 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"24acb3fb4788686e7a0d732ba38c9dcc3fa390170c93afef774c332c4ffb4225"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.762234 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"26f30223dfd78447a69ff61ee06e7e786c94cbf84e82e02030e98543c797a54f"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.762292 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"30ba036cacc3da5cbdff589b1ca7ca9ebcd27b7bc526122fdc96dd84d9924b9b"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.768422 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"03309de27e1d416443ced49271531a500e5163c5ba9e6690f8d7cba69d3b3036"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.768492 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a13ff1c795363efe1d30370a03a6ac54285761effa168080e421c442dfe4fa4b"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.769034 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.777614 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.778059 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.27803379 +0000 UTC m=+151.453769906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.780583 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" event={"ID":"5da410ed-29be-4a38-b486-89f0dd0252ae","Type":"ContainerDied","Data":"0610ca27b55e929da01a90e1b0d9158a5a1e47e4a9997c428e731a5b98ec70b9"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.780625 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0610ca27b55e929da01a90e1b0d9158a5a1e47e4a9997c428e731a5b98ec70b9" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.780726 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330655-552q2" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.788576 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" event={"ID":"d8dd7d8a-f682-4492-8e4d-7318d345a9dd","Type":"ContainerStarted","Data":"0d0ec98803270dd172c8ed827c9207475e84760f06de04c203c520d6769f89fd"} Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.850200 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zlzrn"] Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.850748 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da410ed-29be-4a38-b486-89f0dd0252ae" containerName="collect-profiles" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.850815 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da410ed-29be-4a38-b486-89f0dd0252ae" containerName="collect-profiles" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.850979 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da410ed-29be-4a38-b486-89f0dd0252ae" containerName="collect-profiles" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.851745 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.854541 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.879790 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.882552 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.38253213 +0000 UTC m=+151.558268246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.892124 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlzrn"] Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.983554 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.984394 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-catalog-content\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.984421 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww5dc\" (UniqueName: \"kubernetes.io/projected/536f8e03-4e07-489d-8505-a6fc83e98dc2-kube-api-access-ww5dc\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:34 crc kubenswrapper[4702]: I1007 12:26:34.984463 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-utilities\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:34 crc kubenswrapper[4702]: E1007 12:26:34.984620 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.484597039 +0000 UTC m=+151.660333155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.085956 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.086061 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-catalog-content\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.086085 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww5dc\" (UniqueName: \"kubernetes.io/projected/536f8e03-4e07-489d-8505-a6fc83e98dc2-kube-api-access-ww5dc\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.086482 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.586457182 +0000 UTC m=+151.762193298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.086620 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-catalog-content\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.086701 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-utilities\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.086967 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-utilities\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.115165 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww5dc\" (UniqueName: \"kubernetes.io/projected/536f8e03-4e07-489d-8505-a6fc83e98dc2-kube-api-access-ww5dc\") pod \"certified-operators-zlzrn\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.174391 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.187885 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.188108 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.688077719 +0000 UTC m=+151.863813835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.188266 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.188737 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.688729178 +0000 UTC m=+151.864465294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.197803 4702 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.232702 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bjpd6"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.256626 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.256668 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bjpd6"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.291286 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.291602 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.79158004 +0000 UTC m=+151.967316156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.392805 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.392907 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-catalog-content\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.392943 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbmhj\" (UniqueName: \"kubernetes.io/projected/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-kube-api-access-lbmhj\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.392980 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-utilities\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.393407 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.893389932 +0000 UTC m=+152.069126048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.444720 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bhwkv"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.445800 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.453566 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.455930 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bhwkv"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.470341 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:35 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:35 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:35 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.470406 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.496803 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.497069 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw5hl\" (UniqueName: \"kubernetes.io/projected/55a80c96-1beb-4bc7-b84a-536e43e25b82-kube-api-access-sw5hl\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.497105 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-catalog-content\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.497134 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-catalog-content\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.497153 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbmhj\" (UniqueName: \"kubernetes.io/projected/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-kube-api-access-lbmhj\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.497189 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-utilities\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.497218 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-utilities\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.497346 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:35.997327675 +0000 UTC m=+152.173063791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.498374 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-catalog-content\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.500354 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-utilities\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.524309 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbmhj\" (UniqueName: \"kubernetes.io/projected/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-kube-api-access-lbmhj\") pod \"certified-operators-bjpd6\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.528091 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zlzrn"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.598258 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-utilities\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.598334 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.598375 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw5hl\" (UniqueName: \"kubernetes.io/projected/55a80c96-1beb-4bc7-b84a-536e43e25b82-kube-api-access-sw5hl\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.598401 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-catalog-content\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.598879 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-catalog-content\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.599179 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-utilities\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.599546 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:36.099530408 +0000 UTC m=+152.275266524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.600211 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.623112 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw5hl\" (UniqueName: \"kubernetes.io/projected/55a80c96-1beb-4bc7-b84a-536e43e25b82-kube-api-access-sw5hl\") pod \"community-operators-bhwkv\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.644005 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6fcw2"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.645120 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.666420 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fcw2"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.699825 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.700160 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:36.200122045 +0000 UTC m=+152.375858181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.700302 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-utilities\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.700415 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gv6x\" (UniqueName: \"kubernetes.io/projected/87e06f55-1237-41bb-bf4e-afe9a3342f6a-kube-api-access-9gv6x\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.700518 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.700789 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-catalog-content\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.701052 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:36.201038292 +0000 UTC m=+152.376774408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.765243 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9k26f" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.782602 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.804587 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.804868 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-catalog-content\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.804954 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-utilities\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.804977 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gv6x\" (UniqueName: \"kubernetes.io/projected/87e06f55-1237-41bb-bf4e-afe9a3342f6a-kube-api-access-9gv6x\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.805428 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-catalog-content\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.805547 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 12:26:36.305525071 +0000 UTC m=+152.481261177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.805677 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-utilities\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.845493 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" event={"ID":"d8dd7d8a-f682-4492-8e4d-7318d345a9dd","Type":"ContainerStarted","Data":"62aaaf5dc5641076b8171815e0bbc64f644b11bc683ed4418ab42203fac68176"} Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.845566 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" event={"ID":"d8dd7d8a-f682-4492-8e4d-7318d345a9dd","Type":"ContainerStarted","Data":"f927081e8c7b323c632a3fc0544217cb6b185352ab52e02b01a2504b8ff08c03"} Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.847773 4702 generic.go:334] "Generic (PLEG): container finished" podID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerID="4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297" exitCode=0 Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.848428 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerDied","Data":"4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297"} Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.848460 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerStarted","Data":"09a6cb6c95a08f8533a3fa18301877e6483ed60900fe2beb458e03b941e1a507"} Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.850761 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.857464 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gv6x\" (UniqueName: \"kubernetes.io/projected/87e06f55-1237-41bb-bf4e-afe9a3342f6a-kube-api-access-9gv6x\") pod \"community-operators-6fcw2\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.877003 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-w7whn" podStartSLOduration=11.873877049 podStartE2EDuration="11.873877049s" podCreationTimestamp="2025-10-07 12:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:35.868679379 +0000 UTC m=+152.044415495" watchObservedRunningTime="2025-10-07 12:26:35.873877049 +0000 UTC m=+152.049613165" Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.902315 4702 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-07T12:26:35.19784084Z","Handler":null,"Name":""} Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.913475 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:35 crc kubenswrapper[4702]: E1007 12:26:35.913832 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 12:26:36.413819839 +0000 UTC m=+152.589555955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vw7pc" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.931381 4702 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.931431 4702 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.944052 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bjpd6"] Oct 07 12:26:35 crc kubenswrapper[4702]: I1007 12:26:35.971638 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.015645 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.026386 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.117537 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.132710 4702 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.132784 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.137643 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bhwkv"] Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.170526 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vw7pc\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.233277 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fcw2"] Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.273316 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.273702 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.275109 4702 patch_prober.go:28] interesting pod/console-f9d7485db-q72r8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.275203 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-q72r8" podUID="e95bef62-5bd2-4a64-b1c5-df780f79ac51" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.433996 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.450280 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.450987 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.457061 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.458194 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.463191 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.467793 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:36 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:36 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:36 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.467859 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.509629 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.529815 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.529938 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.633025 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.633512 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.634090 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.651321 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.655312 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.684850 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vw7pc"] Oct 07 12:26:36 crc kubenswrapper[4702]: W1007 12:26:36.695746 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0a452b7_fe35_404d_8864_ff8d6be236da.slice/crio-1758f1ae83b2ed479415c950c78426cf92a51402f394e06252acf0500baacf9e WatchSource:0}: Error finding container 1758f1ae83b2ed479415c950c78426cf92a51402f394e06252acf0500baacf9e: Status 404 returned error can't find the container with id 1758f1ae83b2ed479415c950c78426cf92a51402f394e06252acf0500baacf9e Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.731848 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.731940 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.739151 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r9nf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.739236 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r9nf" podUID="c207de79-b8c1-47f5-8e4b-5f320d760588" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.740110 4702 patch_prober.go:28] interesting pod/apiserver-76f77b778f-mz42v container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]log ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]etcd ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/generic-apiserver-start-informers ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/max-in-flight-filter ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 07 12:26:36 crc kubenswrapper[4702]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 07 12:26:36 crc kubenswrapper[4702]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/project.openshift.io-projectcache ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/openshift.io-startinformers ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 07 12:26:36 crc kubenswrapper[4702]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 07 12:26:36 crc kubenswrapper[4702]: livez check failed Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.740151 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" podUID="e0ec3e7a-b5c0-4245-a6f8-dd783ff20afd" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.740460 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r9nf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.740548 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r9nf" podUID="c207de79-b8c1-47f5-8e4b-5f320d760588" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.820263 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.861167 4702 generic.go:334] "Generic (PLEG): container finished" podID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerID="437e0552801f75247335f171b6a092e4ada63d3c2cc61582fa780f65ad20c4c4" exitCode=0 Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.861240 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerDied","Data":"437e0552801f75247335f171b6a092e4ada63d3c2cc61582fa780f65ad20c4c4"} Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.861369 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerStarted","Data":"4961c4fd94a62de99355d1b9849666f3643215bf846a03fe4c45e7c717ba7d4b"} Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.870510 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" event={"ID":"e0a452b7-fe35-404d-8864-ff8d6be236da","Type":"ContainerStarted","Data":"1758f1ae83b2ed479415c950c78426cf92a51402f394e06252acf0500baacf9e"} Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.873364 4702 generic.go:334] "Generic (PLEG): container finished" podID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerID="d751b3b289a2172e7c7d416b32357119760e87748a9e0381a731f38fa491ca8c" exitCode=0 Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.873598 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhwkv" event={"ID":"55a80c96-1beb-4bc7-b84a-536e43e25b82","Type":"ContainerDied","Data":"d751b3b289a2172e7c7d416b32357119760e87748a9e0381a731f38fa491ca8c"} Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.873889 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhwkv" event={"ID":"55a80c96-1beb-4bc7-b84a-536e43e25b82","Type":"ContainerStarted","Data":"af64a96d48622472acc75d988b476a95207eb7f8383be5632f201c5649efdd53"} Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.878135 4702 generic.go:334] "Generic (PLEG): container finished" podID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerID="0498308f4088121ecc72cd0d04ab911592156202449ae9f6b875ef7b9241f466" exitCode=0 Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.879641 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fcw2" event={"ID":"87e06f55-1237-41bb-bf4e-afe9a3342f6a","Type":"ContainerDied","Data":"0498308f4088121ecc72cd0d04ab911592156202449ae9f6b875ef7b9241f466"} Oct 07 12:26:36 crc kubenswrapper[4702]: I1007 12:26:36.879671 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fcw2" event={"ID":"87e06f55-1237-41bb-bf4e-afe9a3342f6a","Type":"ContainerStarted","Data":"e09de091c378ed44e2705f7d82bfda9ec348f91789d9572552b988fb4aa03de6"} Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.027838 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.027910 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.034524 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9xtg2"] Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.035789 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.037801 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-t56br" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.038994 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.046419 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.051502 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xtg2"] Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.143599 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f49sq\" (UniqueName: \"kubernetes.io/projected/cf22b7ba-6aac-40b0-b25e-72e092658ac6-kube-api-access-f49sq\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.144243 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-catalog-content\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.144403 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-utilities\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.220677 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.246298 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-catalog-content\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.246389 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-utilities\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.246432 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f49sq\" (UniqueName: \"kubernetes.io/projected/cf22b7ba-6aac-40b0-b25e-72e092658ac6-kube-api-access-f49sq\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.247580 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-catalog-content\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.248052 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-utilities\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.266093 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f49sq\" (UniqueName: \"kubernetes.io/projected/cf22b7ba-6aac-40b0-b25e-72e092658ac6-kube-api-access-f49sq\") pod \"redhat-marketplace-9xtg2\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.373784 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.433732 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dlh8f"] Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.435104 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.449327 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlh8f"] Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.464482 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.470468 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:37 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:37 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:37 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.470560 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.541392 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.551348 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7t8c\" (UniqueName: \"kubernetes.io/projected/23e3398a-ac68-4207-9475-3d072d1ce3e4-kube-api-access-g7t8c\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.551398 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-utilities\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.551451 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-catalog-content\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.654289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7t8c\" (UniqueName: \"kubernetes.io/projected/23e3398a-ac68-4207-9475-3d072d1ce3e4-kube-api-access-g7t8c\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.654384 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-utilities\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.654460 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-catalog-content\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.655997 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-catalog-content\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.656336 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-utilities\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.681374 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7t8c\" (UniqueName: \"kubernetes.io/projected/23e3398a-ac68-4207-9475-3d072d1ce3e4-kube-api-access-g7t8c\") pod \"redhat-marketplace-dlh8f\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.711712 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xtg2"] Oct 07 12:26:37 crc kubenswrapper[4702]: W1007 12:26:37.734276 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf22b7ba_6aac_40b0_b25e_72e092658ac6.slice/crio-536e7b5a8d6ba6e42e31a50c4cadf26459735258fd7d7f62b2964487b4af219b WatchSource:0}: Error finding container 536e7b5a8d6ba6e42e31a50c4cadf26459735258fd7d7f62b2964487b4af219b: Status 404 returned error can't find the container with id 536e7b5a8d6ba6e42e31a50c4cadf26459735258fd7d7f62b2964487b4af219b Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.763304 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.892814 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xtg2" event={"ID":"cf22b7ba-6aac-40b0-b25e-72e092658ac6","Type":"ContainerStarted","Data":"536e7b5a8d6ba6e42e31a50c4cadf26459735258fd7d7f62b2964487b4af219b"} Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.897663 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" event={"ID":"e0a452b7-fe35-404d-8864-ff8d6be236da","Type":"ContainerStarted","Data":"6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c"} Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.898731 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.905870 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6beaafb4-6885-4f1f-9f66-1f37326ab8a9","Type":"ContainerStarted","Data":"3d42383d3b6e62c852503926a45a158bccb5daf53b5b31a593c39c012fe6e09b"} Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.905963 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6beaafb4-6885-4f1f-9f66-1f37326ab8a9","Type":"ContainerStarted","Data":"a4a4ba2070c9954816a3d430b769c11be3626af29e776d250f099b5a34d8d781"} Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.912757 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vxk4" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.931819 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" podStartSLOduration=128.931795713 podStartE2EDuration="2m8.931795713s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:37.923734701 +0000 UTC m=+154.099470827" watchObservedRunningTime="2025-10-07 12:26:37.931795713 +0000 UTC m=+154.107531819" Oct 07 12:26:37 crc kubenswrapper[4702]: I1007 12:26:37.944233 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.9441771 podStartE2EDuration="1.9441771s" podCreationTimestamp="2025-10-07 12:26:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:37.94072184 +0000 UTC m=+154.116457956" watchObservedRunningTime="2025-10-07 12:26:37.9441771 +0000 UTC m=+154.119913216" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.207712 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlh8f"] Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.237413 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8zjkc"] Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.238825 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.242382 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.248587 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zjkc"] Oct 07 12:26:38 crc kubenswrapper[4702]: W1007 12:26:38.269364 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23e3398a_ac68_4207_9475_3d072d1ce3e4.slice/crio-9dd7924696842bf6dd552d4890c01cf7a626246f1e66c9484fe8a29f5c4c3aba WatchSource:0}: Error finding container 9dd7924696842bf6dd552d4890c01cf7a626246f1e66c9484fe8a29f5c4c3aba: Status 404 returned error can't find the container with id 9dd7924696842bf6dd552d4890c01cf7a626246f1e66c9484fe8a29f5c4c3aba Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.272022 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb9hf\" (UniqueName: \"kubernetes.io/projected/988399d3-c52f-450e-a74f-3fd6bbf03c1f-kube-api-access-cb9hf\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.272085 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-catalog-content\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.272253 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-utilities\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.373860 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-utilities\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.374005 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb9hf\" (UniqueName: \"kubernetes.io/projected/988399d3-c52f-450e-a74f-3fd6bbf03c1f-kube-api-access-cb9hf\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.374061 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-catalog-content\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.374678 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-utilities\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.374727 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-catalog-content\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.396287 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb9hf\" (UniqueName: \"kubernetes.io/projected/988399d3-c52f-450e-a74f-3fd6bbf03c1f-kube-api-access-cb9hf\") pod \"redhat-operators-8zjkc\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.467613 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:38 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:38 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:38 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.467697 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.574267 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.632514 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f5mxz"] Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.634512 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.692239 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5mxz"] Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.785400 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-catalog-content\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.785475 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-utilities\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.785499 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd2kf\" (UniqueName: \"kubernetes.io/projected/2247b061-f50a-40d4-bbab-b25a6284decf-kube-api-access-kd2kf\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.880568 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zjkc"] Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.887373 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-utilities\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.887431 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd2kf\" (UniqueName: \"kubernetes.io/projected/2247b061-f50a-40d4-bbab-b25a6284decf-kube-api-access-kd2kf\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.887505 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-catalog-content\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.888607 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-utilities\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.889096 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-catalog-content\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: W1007 12:26:38.896156 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod988399d3_c52f_450e_a74f_3fd6bbf03c1f.slice/crio-3ce2e0705658d2f67cb6d23c85cdfd07ba6f4e3c231250afa29be001692ff98c WatchSource:0}: Error finding container 3ce2e0705658d2f67cb6d23c85cdfd07ba6f4e3c231250afa29be001692ff98c: Status 404 returned error can't find the container with id 3ce2e0705658d2f67cb6d23c85cdfd07ba6f4e3c231250afa29be001692ff98c Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.912411 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd2kf\" (UniqueName: \"kubernetes.io/projected/2247b061-f50a-40d4-bbab-b25a6284decf-kube-api-access-kd2kf\") pod \"redhat-operators-f5mxz\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.927129 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjkc" event={"ID":"988399d3-c52f-450e-a74f-3fd6bbf03c1f","Type":"ContainerStarted","Data":"3ce2e0705658d2f67cb6d23c85cdfd07ba6f4e3c231250afa29be001692ff98c"} Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.940756 4702 generic.go:334] "Generic (PLEG): container finished" podID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerID="6eb50bdd69ea6a3417672c0f997ca03d566460d0ad375d18f641d95ca21857f2" exitCode=0 Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.940844 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xtg2" event={"ID":"cf22b7ba-6aac-40b0-b25e-72e092658ac6","Type":"ContainerDied","Data":"6eb50bdd69ea6a3417672c0f997ca03d566460d0ad375d18f641d95ca21857f2"} Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.952887 4702 generic.go:334] "Generic (PLEG): container finished" podID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerID="e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c" exitCode=0 Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.953221 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerDied","Data":"e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c"} Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.953298 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerStarted","Data":"9dd7924696842bf6dd552d4890c01cf7a626246f1e66c9484fe8a29f5c4c3aba"} Oct 07 12:26:38 crc kubenswrapper[4702]: I1007 12:26:38.999586 4702 generic.go:334] "Generic (PLEG): container finished" podID="6beaafb4-6885-4f1f-9f66-1f37326ab8a9" containerID="3d42383d3b6e62c852503926a45a158bccb5daf53b5b31a593c39c012fe6e09b" exitCode=0 Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.000875 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6beaafb4-6885-4f1f-9f66-1f37326ab8a9","Type":"ContainerDied","Data":"3d42383d3b6e62c852503926a45a158bccb5daf53b5b31a593c39c012fe6e09b"} Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.011305 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.264835 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.265307 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.365427 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5mxz"] Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.468098 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:39 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:39 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:39 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:39 crc kubenswrapper[4702]: I1007 12:26:39.468217 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.007720 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerStarted","Data":"e6066b356a4ac2cae9d316c50f01a5ef62d5456d64a5cf4db27efb77f285a58d"} Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.011025 4702 generic.go:334] "Generic (PLEG): container finished" podID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerID="98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f" exitCode=0 Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.011284 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjkc" event={"ID":"988399d3-c52f-450e-a74f-3fd6bbf03c1f","Type":"ContainerDied","Data":"98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f"} Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.352047 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.368161 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 12:26:40 crc kubenswrapper[4702]: E1007 12:26:40.369409 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6beaafb4-6885-4f1f-9f66-1f37326ab8a9" containerName="pruner" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.369426 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="6beaafb4-6885-4f1f-9f66-1f37326ab8a9" containerName="pruner" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.369563 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="6beaafb4-6885-4f1f-9f66-1f37326ab8a9" containerName="pruner" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.372610 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.375795 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.376394 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.377189 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.467449 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:40 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:40 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:40 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.467513 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.514035 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kube-api-access\") pod \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.514289 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kubelet-dir\") pod \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\" (UID: \"6beaafb4-6885-4f1f-9f66-1f37326ab8a9\") " Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.514461 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6beaafb4-6885-4f1f-9f66-1f37326ab8a9" (UID: "6beaafb4-6885-4f1f-9f66-1f37326ab8a9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.514535 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.514760 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.515072 4702 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.521635 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6beaafb4-6885-4f1f-9f66-1f37326ab8a9" (UID: "6beaafb4-6885-4f1f-9f66-1f37326ab8a9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.617432 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.617559 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.617625 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6beaafb4-6885-4f1f-9f66-1f37326ab8a9-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.617632 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.635812 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.697449 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:40 crc kubenswrapper[4702]: I1007 12:26:40.967815 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.025299 4702 generic.go:334] "Generic (PLEG): container finished" podID="2247b061-f50a-40d4-bbab-b25a6284decf" containerID="986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd" exitCode=0 Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.025388 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerDied","Data":"986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd"} Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.028482 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6beaafb4-6885-4f1f-9f66-1f37326ab8a9","Type":"ContainerDied","Data":"a4a4ba2070c9954816a3d430b769c11be3626af29e776d250f099b5a34d8d781"} Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.028544 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4a4ba2070c9954816a3d430b769c11be3626af29e776d250f099b5a34d8d781" Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.028605 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.467946 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:41 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:41 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:41 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.468283 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.808498 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:41 crc kubenswrapper[4702]: I1007 12:26:41.813413 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-mz42v" Oct 07 12:26:42 crc kubenswrapper[4702]: I1007 12:26:42.050228 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed634baa-b36c-4a65-b8eb-84cf925a17f2","Type":"ContainerStarted","Data":"e5be2dac0c3bc6f03e1c0ce4741568ef1137b6328cd1107960471d82801e0f06"} Oct 07 12:26:42 crc kubenswrapper[4702]: I1007 12:26:42.050296 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed634baa-b36c-4a65-b8eb-84cf925a17f2","Type":"ContainerStarted","Data":"eab0ab7a67fa0a4f7eabc6c3264abf233cff3c2817a53f589a0f30b648948197"} Oct 07 12:26:42 crc kubenswrapper[4702]: I1007 12:26:42.466922 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:42 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:42 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:42 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:42 crc kubenswrapper[4702]: I1007 12:26:42.466993 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:42 crc kubenswrapper[4702]: I1007 12:26:42.582680 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-c5n5d" Oct 07 12:26:43 crc kubenswrapper[4702]: I1007 12:26:43.072329 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.072300562 podStartE2EDuration="3.072300562s" podCreationTimestamp="2025-10-07 12:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:26:43.071153819 +0000 UTC m=+159.246889935" watchObservedRunningTime="2025-10-07 12:26:43.072300562 +0000 UTC m=+159.248036678" Oct 07 12:26:43 crc kubenswrapper[4702]: I1007 12:26:43.467736 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:43 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:43 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:43 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:43 crc kubenswrapper[4702]: I1007 12:26:43.467865 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:44 crc kubenswrapper[4702]: I1007 12:26:44.066982 4702 generic.go:334] "Generic (PLEG): container finished" podID="ed634baa-b36c-4a65-b8eb-84cf925a17f2" containerID="e5be2dac0c3bc6f03e1c0ce4741568ef1137b6328cd1107960471d82801e0f06" exitCode=0 Oct 07 12:26:44 crc kubenswrapper[4702]: I1007 12:26:44.067090 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed634baa-b36c-4a65-b8eb-84cf925a17f2","Type":"ContainerDied","Data":"e5be2dac0c3bc6f03e1c0ce4741568ef1137b6328cd1107960471d82801e0f06"} Oct 07 12:26:44 crc kubenswrapper[4702]: I1007 12:26:44.468422 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:44 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:44 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:44 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:44 crc kubenswrapper[4702]: I1007 12:26:44.468510 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:45 crc kubenswrapper[4702]: I1007 12:26:45.467132 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:45 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:45 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:45 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:45 crc kubenswrapper[4702]: I1007 12:26:45.467570 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.274784 4702 patch_prober.go:28] interesting pod/console-f9d7485db-q72r8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.274946 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-q72r8" podUID="e95bef62-5bd2-4a64-b1c5-df780f79ac51" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.468711 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:46 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:46 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:46 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.469268 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.737765 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r9nf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.737825 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r9nf" podUID="c207de79-b8c1-47f5-8e4b-5f320d760588" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.737862 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r9nf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 07 12:26:46 crc kubenswrapper[4702]: I1007 12:26:46.737885 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r9nf" podUID="c207de79-b8c1-47f5-8e4b-5f320d760588" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 07 12:26:47 crc kubenswrapper[4702]: I1007 12:26:47.466866 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:47 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:47 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:47 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:47 crc kubenswrapper[4702]: I1007 12:26:47.466944 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:48 crc kubenswrapper[4702]: I1007 12:26:48.466856 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:48 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:48 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:48 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:48 crc kubenswrapper[4702]: I1007 12:26:48.466974 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.444258 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.471512 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:49 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:49 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:49 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.471602 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.551440 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kube-api-access\") pod \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.551753 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kubelet-dir\") pod \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\" (UID: \"ed634baa-b36c-4a65-b8eb-84cf925a17f2\") " Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.552017 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ed634baa-b36c-4a65-b8eb-84cf925a17f2" (UID: "ed634baa-b36c-4a65-b8eb-84cf925a17f2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.552763 4702 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.558259 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ed634baa-b36c-4a65-b8eb-84cf925a17f2" (UID: "ed634baa-b36c-4a65-b8eb-84cf925a17f2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:26:49 crc kubenswrapper[4702]: I1007 12:26:49.654582 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed634baa-b36c-4a65-b8eb-84cf925a17f2-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 12:26:50 crc kubenswrapper[4702]: I1007 12:26:50.106739 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed634baa-b36c-4a65-b8eb-84cf925a17f2","Type":"ContainerDied","Data":"eab0ab7a67fa0a4f7eabc6c3264abf233cff3c2817a53f589a0f30b648948197"} Oct 07 12:26:50 crc kubenswrapper[4702]: I1007 12:26:50.106792 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eab0ab7a67fa0a4f7eabc6c3264abf233cff3c2817a53f589a0f30b648948197" Oct 07 12:26:50 crc kubenswrapper[4702]: I1007 12:26:50.107380 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 12:26:50 crc kubenswrapper[4702]: I1007 12:26:50.469432 4702 patch_prober.go:28] interesting pod/router-default-5444994796-n66bc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 12:26:50 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Oct 07 12:26:50 crc kubenswrapper[4702]: [+]process-running ok Oct 07 12:26:50 crc kubenswrapper[4702]: healthz check failed Oct 07 12:26:50 crc kubenswrapper[4702]: I1007 12:26:50.469585 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n66bc" podUID="3602b00e-3c66-4dd4-94fb-cc4421ea7566" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 12:26:51 crc kubenswrapper[4702]: I1007 12:26:51.467755 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:51 crc kubenswrapper[4702]: I1007 12:26:51.470694 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-n66bc" Oct 07 12:26:51 crc kubenswrapper[4702]: I1007 12:26:51.991061 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:51 crc kubenswrapper[4702]: I1007 12:26:51.996252 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fb0cea3-d8a9-4738-abf3-47eea11213d7-metrics-certs\") pod \"network-metrics-daemon-znndd\" (UID: \"8fb0cea3-d8a9-4738-abf3-47eea11213d7\") " pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:52 crc kubenswrapper[4702]: I1007 12:26:52.257569 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-znndd" Oct 07 12:26:56 crc kubenswrapper[4702]: I1007 12:26:56.313293 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:56 crc kubenswrapper[4702]: I1007 12:26:56.319645 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-q72r8" Oct 07 12:26:56 crc kubenswrapper[4702]: I1007 12:26:56.439718 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:26:56 crc kubenswrapper[4702]: I1007 12:26:56.749015 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9r9nf" Oct 07 12:27:07 crc kubenswrapper[4702]: I1007 12:27:07.525674 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l9r92" Oct 07 12:27:09 crc kubenswrapper[4702]: I1007 12:27:09.262811 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:27:09 crc kubenswrapper[4702]: I1007 12:27:09.262921 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:27:12 crc kubenswrapper[4702]: I1007 12:27:12.979108 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.421804 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.422064 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw5hl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-bhwkv_openshift-marketplace(55a80c96-1beb-4bc7-b84a-536e43e25b82): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.423448 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-bhwkv" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.540396 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.540939 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6fcw2_openshift-marketplace(87e06f55-1237-41bb-bf4e-afe9a3342f6a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.542190 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6fcw2" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.621203 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.623094 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbmhj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bjpd6_openshift-marketplace(27b2dc0e-3eab-47f5-a64d-b6b74f720dcd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.624408 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-bjpd6" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.652264 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.652764 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ww5dc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zlzrn_openshift-marketplace(536f8e03-4e07-489d-8505-a6fc83e98dc2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:19 crc kubenswrapper[4702]: E1007 12:27:19.653919 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zlzrn" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.446350 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6fcw2" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.446363 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bjpd6" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.446376 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-bhwkv" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.446492 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zlzrn" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.996046 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:afa798da1eea334bab3cb1e14451ff84f98d35b436cdc4b408b46e289e4e2bc2: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:afa798da1eea334bab3cb1e14451ff84f98d35b436cdc4b408b46e289e4e2bc2\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.996496 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f49sq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9xtg2_openshift-marketplace(cf22b7ba-6aac-40b0-b25e-72e092658ac6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:afa798da1eea334bab3cb1e14451ff84f98d35b436cdc4b408b46e289e4e2bc2: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:afa798da1eea334bab3cb1e14451ff84f98d35b436cdc4b408b46e289e4e2bc2\": context canceled" logger="UnhandledError" Oct 07 12:27:20 crc kubenswrapper[4702]: E1007 12:27:20.997688 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:afa798da1eea334bab3cb1e14451ff84f98d35b436cdc4b408b46e289e4e2bc2: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:afa798da1eea334bab3cb1e14451ff84f98d35b436cdc4b408b46e289e4e2bc2\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-9xtg2" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.120570 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9xtg2" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.143474 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.143689 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kd2kf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-f5mxz_openshift-marketplace(2247b061-f50a-40d4-bbab-b25a6284decf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.145064 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-f5mxz" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.150615 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.150818 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cb9hf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-8zjkc_openshift-marketplace(988399d3-c52f-450e-a74f-3fd6bbf03c1f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:24 crc kubenswrapper[4702]: E1007 12:27:24.152145 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-8zjkc" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" Oct 07 12:27:26 crc kubenswrapper[4702]: E1007 12:27:26.818689 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-8zjkc" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" Oct 07 12:27:27 crc kubenswrapper[4702]: E1007 12:27:27.442971 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-f5mxz" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" Oct 07 12:27:27 crc kubenswrapper[4702]: E1007 12:27:27.506045 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 12:27:27 crc kubenswrapper[4702]: E1007 12:27:27.506349 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g7t8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dlh8f_openshift-marketplace(23e3398a-ac68-4207-9475-3d072d1ce3e4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:27:27 crc kubenswrapper[4702]: E1007 12:27:27.508094 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dlh8f" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" Oct 07 12:27:27 crc kubenswrapper[4702]: I1007 12:27:27.850061 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-znndd"] Oct 07 12:27:27 crc kubenswrapper[4702]: W1007 12:27:27.859316 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fb0cea3_d8a9_4738_abf3_47eea11213d7.slice/crio-bdb71c1b3bc4314af3f697ffa3f7779f36620edf5ec224af5bf4a0dbc2425a9a WatchSource:0}: Error finding container bdb71c1b3bc4314af3f697ffa3f7779f36620edf5ec224af5bf4a0dbc2425a9a: Status 404 returned error can't find the container with id bdb71c1b3bc4314af3f697ffa3f7779f36620edf5ec224af5bf4a0dbc2425a9a Oct 07 12:27:28 crc kubenswrapper[4702]: I1007 12:27:28.339130 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-znndd" event={"ID":"8fb0cea3-d8a9-4738-abf3-47eea11213d7","Type":"ContainerStarted","Data":"bfb6b0e78857cbf8d27778428272e15cf28a315e7025ea17a12574585f4335e8"} Oct 07 12:27:28 crc kubenswrapper[4702]: I1007 12:27:28.339697 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-znndd" event={"ID":"8fb0cea3-d8a9-4738-abf3-47eea11213d7","Type":"ContainerStarted","Data":"8df04081ad4219c95cc30f29e6b76ef1b46e6cf3e84155635f13e317604c0c8b"} Oct 07 12:27:28 crc kubenswrapper[4702]: I1007 12:27:28.339715 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-znndd" event={"ID":"8fb0cea3-d8a9-4738-abf3-47eea11213d7","Type":"ContainerStarted","Data":"bdb71c1b3bc4314af3f697ffa3f7779f36620edf5ec224af5bf4a0dbc2425a9a"} Oct 07 12:27:28 crc kubenswrapper[4702]: E1007 12:27:28.341056 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dlh8f" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" Oct 07 12:27:28 crc kubenswrapper[4702]: I1007 12:27:28.376766 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-znndd" podStartSLOduration=179.37672801 podStartE2EDuration="2m59.37672801s" podCreationTimestamp="2025-10-07 12:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:27:28.37209439 +0000 UTC m=+204.547830506" watchObservedRunningTime="2025-10-07 12:27:28.37672801 +0000 UTC m=+204.552464166" Oct 07 12:27:35 crc kubenswrapper[4702]: I1007 12:27:35.386029 4702 generic.go:334] "Generic (PLEG): container finished" podID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerID="2ebf64ebe35e952b79b4afce0299879dec9440d8d6fc551cbc7e839b7baf9f31" exitCode=0 Oct 07 12:27:35 crc kubenswrapper[4702]: I1007 12:27:35.386127 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fcw2" event={"ID":"87e06f55-1237-41bb-bf4e-afe9a3342f6a","Type":"ContainerDied","Data":"2ebf64ebe35e952b79b4afce0299879dec9440d8d6fc551cbc7e839b7baf9f31"} Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.401639 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerStarted","Data":"5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4"} Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.407472 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fcw2" event={"ID":"87e06f55-1237-41bb-bf4e-afe9a3342f6a","Type":"ContainerStarted","Data":"9d7cedfee3f699f1f6d088703fce49b46db63949dfb45288e665a3d8f97e4403"} Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.411619 4702 generic.go:334] "Generic (PLEG): container finished" podID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerID="f34ea6b84816395c80aff712f49e84ab6bb5466373cbbd3835a1cd2abab8a1f4" exitCode=0 Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.411732 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xtg2" event={"ID":"cf22b7ba-6aac-40b0-b25e-72e092658ac6","Type":"ContainerDied","Data":"f34ea6b84816395c80aff712f49e84ab6bb5466373cbbd3835a1cd2abab8a1f4"} Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.421658 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerStarted","Data":"6e2068fa7c744011518ee06e1c7788d82842e83d02e5c7619400345f5c5984a5"} Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.432467 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhwkv" event={"ID":"55a80c96-1beb-4bc7-b84a-536e43e25b82","Type":"ContainerStarted","Data":"4a82f11900f5e9f77cb3cb83bf699e4f436a94e6da066831eab0cf924ce85036"} Oct 07 12:27:37 crc kubenswrapper[4702]: I1007 12:27:37.475944 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6fcw2" podStartSLOduration=2.336795403 podStartE2EDuration="1m2.475915963s" podCreationTimestamp="2025-10-07 12:26:35 +0000 UTC" firstStartedPulling="2025-10-07 12:26:36.885258926 +0000 UTC m=+153.060995042" lastFinishedPulling="2025-10-07 12:27:37.024379476 +0000 UTC m=+213.200115602" observedRunningTime="2025-10-07 12:27:37.470236332 +0000 UTC m=+213.645972448" watchObservedRunningTime="2025-10-07 12:27:37.475915963 +0000 UTC m=+213.651652079" Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.442113 4702 generic.go:334] "Generic (PLEG): container finished" podID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerID="6e2068fa7c744011518ee06e1c7788d82842e83d02e5c7619400345f5c5984a5" exitCode=0 Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.442193 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerDied","Data":"6e2068fa7c744011518ee06e1c7788d82842e83d02e5c7619400345f5c5984a5"} Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.446647 4702 generic.go:334] "Generic (PLEG): container finished" podID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerID="4a82f11900f5e9f77cb3cb83bf699e4f436a94e6da066831eab0cf924ce85036" exitCode=0 Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.446752 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhwkv" event={"ID":"55a80c96-1beb-4bc7-b84a-536e43e25b82","Type":"ContainerDied","Data":"4a82f11900f5e9f77cb3cb83bf699e4f436a94e6da066831eab0cf924ce85036"} Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.449805 4702 generic.go:334] "Generic (PLEG): container finished" podID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerID="5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4" exitCode=0 Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.449892 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerDied","Data":"5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4"} Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.453347 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xtg2" event={"ID":"cf22b7ba-6aac-40b0-b25e-72e092658ac6","Type":"ContainerStarted","Data":"ca1d8e878af469855bd808fc19ba8d9c1342a05e288cd0bcf81a8de99ac33070"} Oct 07 12:27:38 crc kubenswrapper[4702]: I1007 12:27:38.534758 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9xtg2" podStartSLOduration=2.567375303 podStartE2EDuration="1m1.534736165s" podCreationTimestamp="2025-10-07 12:26:37 +0000 UTC" firstStartedPulling="2025-10-07 12:26:38.952121618 +0000 UTC m=+155.127857734" lastFinishedPulling="2025-10-07 12:27:37.91948244 +0000 UTC m=+214.095218596" observedRunningTime="2025-10-07 12:27:38.533515028 +0000 UTC m=+214.709251144" watchObservedRunningTime="2025-10-07 12:27:38.534736165 +0000 UTC m=+214.710472281" Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.262999 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.263967 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.264561 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.265400 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.265659 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342" gracePeriod=600 Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.469314 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhwkv" event={"ID":"55a80c96-1beb-4bc7-b84a-536e43e25b82","Type":"ContainerStarted","Data":"5dbbbb8db5d824382b20725348694e995a54c726122b01ca035ce53ca179cd12"} Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.474930 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerStarted","Data":"e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d"} Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.476761 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342" exitCode=0 Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.476823 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342"} Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.478791 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerStarted","Data":"c9a131333dee0061ff73ff46ba2e3acc752075dc61525bca1354e66e6ff49a91"} Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.501842 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bhwkv" podStartSLOduration=2.397129242 podStartE2EDuration="1m4.501809496s" podCreationTimestamp="2025-10-07 12:26:35 +0000 UTC" firstStartedPulling="2025-10-07 12:26:36.875492824 +0000 UTC m=+153.051228940" lastFinishedPulling="2025-10-07 12:27:38.980173068 +0000 UTC m=+215.155909194" observedRunningTime="2025-10-07 12:27:39.486908457 +0000 UTC m=+215.662644593" watchObservedRunningTime="2025-10-07 12:27:39.501809496 +0000 UTC m=+215.677545612" Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.537817 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zlzrn" podStartSLOduration=2.418743287 podStartE2EDuration="1m5.537791439s" podCreationTimestamp="2025-10-07 12:26:34 +0000 UTC" firstStartedPulling="2025-10-07 12:26:35.850494086 +0000 UTC m=+152.026230202" lastFinishedPulling="2025-10-07 12:27:38.969542238 +0000 UTC m=+215.145278354" observedRunningTime="2025-10-07 12:27:39.537769638 +0000 UTC m=+215.713505764" watchObservedRunningTime="2025-10-07 12:27:39.537791439 +0000 UTC m=+215.713527555" Oct 07 12:27:39 crc kubenswrapper[4702]: I1007 12:27:39.542872 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bjpd6" podStartSLOduration=2.309521097 podStartE2EDuration="1m4.542851542s" podCreationTimestamp="2025-10-07 12:26:35 +0000 UTC" firstStartedPulling="2025-10-07 12:26:36.865518277 +0000 UTC m=+153.041254393" lastFinishedPulling="2025-10-07 12:27:39.098848722 +0000 UTC m=+215.274584838" observedRunningTime="2025-10-07 12:27:39.511612701 +0000 UTC m=+215.687348827" watchObservedRunningTime="2025-10-07 12:27:39.542851542 +0000 UTC m=+215.718587658" Oct 07 12:27:40 crc kubenswrapper[4702]: I1007 12:27:40.494428 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerStarted","Data":"4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc"} Oct 07 12:27:40 crc kubenswrapper[4702]: I1007 12:27:40.498213 4702 generic.go:334] "Generic (PLEG): container finished" podID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerID="99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426" exitCode=0 Oct 07 12:27:40 crc kubenswrapper[4702]: I1007 12:27:40.498306 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjkc" event={"ID":"988399d3-c52f-450e-a74f-3fd6bbf03c1f","Type":"ContainerDied","Data":"99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426"} Oct 07 12:27:40 crc kubenswrapper[4702]: I1007 12:27:40.502770 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"446febc400f1513a5b71f983f1776c7a64b2c92e3c79409f2d677ce29b1b6901"} Oct 07 12:27:41 crc kubenswrapper[4702]: I1007 12:27:41.511072 4702 generic.go:334] "Generic (PLEG): container finished" podID="2247b061-f50a-40d4-bbab-b25a6284decf" containerID="4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc" exitCode=0 Oct 07 12:27:41 crc kubenswrapper[4702]: I1007 12:27:41.512389 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerDied","Data":"4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc"} Oct 07 12:27:42 crc kubenswrapper[4702]: I1007 12:27:42.521267 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjkc" event={"ID":"988399d3-c52f-450e-a74f-3fd6bbf03c1f","Type":"ContainerStarted","Data":"0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3"} Oct 07 12:27:42 crc kubenswrapper[4702]: I1007 12:27:42.545475 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8zjkc" podStartSLOduration=3.10563768 podStartE2EDuration="1m4.545449665s" podCreationTimestamp="2025-10-07 12:26:38 +0000 UTC" firstStartedPulling="2025-10-07 12:26:40.012876986 +0000 UTC m=+156.188613102" lastFinishedPulling="2025-10-07 12:27:41.452688971 +0000 UTC m=+217.628425087" observedRunningTime="2025-10-07 12:27:42.544108894 +0000 UTC m=+218.719845010" watchObservedRunningTime="2025-10-07 12:27:42.545449665 +0000 UTC m=+218.721185781" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.175494 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.175909 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.601768 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.602437 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.783920 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.783967 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.972733 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:27:45 crc kubenswrapper[4702]: I1007 12:27:45.972820 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.384762 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.385076 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.385148 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.393821 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.430913 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.593203 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.594529 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:27:46 crc kubenswrapper[4702]: I1007 12:27:46.602800 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:27:47 crc kubenswrapper[4702]: I1007 12:27:47.374780 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:27:47 crc kubenswrapper[4702]: I1007 12:27:47.375324 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:27:47 crc kubenswrapper[4702]: I1007 12:27:47.431435 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:27:47 crc kubenswrapper[4702]: I1007 12:27:47.558977 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerStarted","Data":"8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f"} Oct 07 12:27:47 crc kubenswrapper[4702]: I1007 12:27:47.581679 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f5mxz" podStartSLOduration=4.486962381 podStartE2EDuration="1m9.581659694s" podCreationTimestamp="2025-10-07 12:26:38 +0000 UTC" firstStartedPulling="2025-10-07 12:26:41.035853505 +0000 UTC m=+157.211589621" lastFinishedPulling="2025-10-07 12:27:46.130550818 +0000 UTC m=+222.306286934" observedRunningTime="2025-10-07 12:27:47.581094427 +0000 UTC m=+223.756830553" watchObservedRunningTime="2025-10-07 12:27:47.581659694 +0000 UTC m=+223.757395810" Oct 07 12:27:47 crc kubenswrapper[4702]: I1007 12:27:47.603367 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:27:48 crc kubenswrapper[4702]: I1007 12:27:48.575710 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:27:48 crc kubenswrapper[4702]: I1007 12:27:48.575989 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:27:48 crc kubenswrapper[4702]: I1007 12:27:48.638032 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:27:48 crc kubenswrapper[4702]: I1007 12:27:48.685847 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bjpd6"] Oct 07 12:27:48 crc kubenswrapper[4702]: I1007 12:27:48.686372 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bjpd6" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="registry-server" containerID="cri-o://c9a131333dee0061ff73ff46ba2e3acc752075dc61525bca1354e66e6ff49a91" gracePeriod=2 Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.012168 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.012233 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.285962 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fcw2"] Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.286675 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6fcw2" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="registry-server" containerID="cri-o://9d7cedfee3f699f1f6d088703fce49b46db63949dfb45288e665a3d8f97e4403" gracePeriod=2 Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.573375 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerStarted","Data":"983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad"} Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.575602 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerDied","Data":"c9a131333dee0061ff73ff46ba2e3acc752075dc61525bca1354e66e6ff49a91"} Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.578623 4702 generic.go:334] "Generic (PLEG): container finished" podID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerID="c9a131333dee0061ff73ff46ba2e3acc752075dc61525bca1354e66e6ff49a91" exitCode=0 Oct 07 12:27:49 crc kubenswrapper[4702]: I1007 12:27:49.622565 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.071299 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f5mxz" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="registry-server" probeResult="failure" output=< Oct 07 12:27:50 crc kubenswrapper[4702]: timeout: failed to connect service ":50051" within 1s Oct 07 12:27:50 crc kubenswrapper[4702]: > Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.557927 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.601710 4702 generic.go:334] "Generic (PLEG): container finished" podID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerID="983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad" exitCode=0 Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.601858 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerDied","Data":"983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad"} Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.606135 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjpd6" event={"ID":"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd","Type":"ContainerDied","Data":"4961c4fd94a62de99355d1b9849666f3643215bf846a03fe4c45e7c717ba7d4b"} Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.606189 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjpd6" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.606208 4702 scope.go:117] "RemoveContainer" containerID="c9a131333dee0061ff73ff46ba2e3acc752075dc61525bca1354e66e6ff49a91" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.634535 4702 generic.go:334] "Generic (PLEG): container finished" podID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerID="9d7cedfee3f699f1f6d088703fce49b46db63949dfb45288e665a3d8f97e4403" exitCode=0 Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.634596 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fcw2" event={"ID":"87e06f55-1237-41bb-bf4e-afe9a3342f6a","Type":"ContainerDied","Data":"9d7cedfee3f699f1f6d088703fce49b46db63949dfb45288e665a3d8f97e4403"} Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.654129 4702 scope.go:117] "RemoveContainer" containerID="6e2068fa7c744011518ee06e1c7788d82842e83d02e5c7619400345f5c5984a5" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.677947 4702 scope.go:117] "RemoveContainer" containerID="437e0552801f75247335f171b6a092e4ada63d3c2cc61582fa780f65ad20c4c4" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.701606 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-catalog-content\") pod \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.701699 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-utilities\") pod \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.701856 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbmhj\" (UniqueName: \"kubernetes.io/projected/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-kube-api-access-lbmhj\") pod \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\" (UID: \"27b2dc0e-3eab-47f5-a64d-b6b74f720dcd\") " Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.706580 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-utilities" (OuterVolumeSpecName: "utilities") pod "27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" (UID: "27b2dc0e-3eab-47f5-a64d-b6b74f720dcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.713179 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-kube-api-access-lbmhj" (OuterVolumeSpecName: "kube-api-access-lbmhj") pod "27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" (UID: "27b2dc0e-3eab-47f5-a64d-b6b74f720dcd"). InnerVolumeSpecName "kube-api-access-lbmhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.803449 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.803502 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbmhj\" (UniqueName: \"kubernetes.io/projected/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-kube-api-access-lbmhj\") on node \"crc\" DevicePath \"\"" Oct 07 12:27:50 crc kubenswrapper[4702]: I1007 12:27:50.944458 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" (UID: "27b2dc0e-3eab-47f5-a64d-b6b74f720dcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.006010 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.176746 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.238331 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bjpd6"] Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.241396 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bjpd6"] Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.311869 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gv6x\" (UniqueName: \"kubernetes.io/projected/87e06f55-1237-41bb-bf4e-afe9a3342f6a-kube-api-access-9gv6x\") pod \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.312032 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-utilities\") pod \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.312083 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-catalog-content\") pod \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\" (UID: \"87e06f55-1237-41bb-bf4e-afe9a3342f6a\") " Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.313092 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-utilities" (OuterVolumeSpecName: "utilities") pod "87e06f55-1237-41bb-bf4e-afe9a3342f6a" (UID: "87e06f55-1237-41bb-bf4e-afe9a3342f6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.315104 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e06f55-1237-41bb-bf4e-afe9a3342f6a-kube-api-access-9gv6x" (OuterVolumeSpecName: "kube-api-access-9gv6x") pod "87e06f55-1237-41bb-bf4e-afe9a3342f6a" (UID: "87e06f55-1237-41bb-bf4e-afe9a3342f6a"). InnerVolumeSpecName "kube-api-access-9gv6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.382852 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87e06f55-1237-41bb-bf4e-afe9a3342f6a" (UID: "87e06f55-1237-41bb-bf4e-afe9a3342f6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.414398 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.414447 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e06f55-1237-41bb-bf4e-afe9a3342f6a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.414459 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gv6x\" (UniqueName: \"kubernetes.io/projected/87e06f55-1237-41bb-bf4e-afe9a3342f6a-kube-api-access-9gv6x\") on node \"crc\" DevicePath \"\"" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.643526 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fcw2" event={"ID":"87e06f55-1237-41bb-bf4e-afe9a3342f6a","Type":"ContainerDied","Data":"e09de091c378ed44e2705f7d82bfda9ec348f91789d9572552b988fb4aa03de6"} Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.644132 4702 scope.go:117] "RemoveContainer" containerID="9d7cedfee3f699f1f6d088703fce49b46db63949dfb45288e665a3d8f97e4403" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.643594 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fcw2" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.683571 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fcw2"] Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.684398 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6fcw2"] Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.689640 4702 scope.go:117] "RemoveContainer" containerID="2ebf64ebe35e952b79b4afce0299879dec9440d8d6fc551cbc7e839b7baf9f31" Oct 07 12:27:51 crc kubenswrapper[4702]: I1007 12:27:51.707018 4702 scope.go:117] "RemoveContainer" containerID="0498308f4088121ecc72cd0d04ab911592156202449ae9f6b875ef7b9241f466" Oct 07 12:27:52 crc kubenswrapper[4702]: I1007 12:27:52.647643 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" path="/var/lib/kubelet/pods/27b2dc0e-3eab-47f5-a64d-b6b74f720dcd/volumes" Oct 07 12:27:52 crc kubenswrapper[4702]: I1007 12:27:52.648344 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" path="/var/lib/kubelet/pods/87e06f55-1237-41bb-bf4e-afe9a3342f6a/volumes" Oct 07 12:27:54 crc kubenswrapper[4702]: I1007 12:27:54.681802 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerStarted","Data":"b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb"} Oct 07 12:27:57 crc kubenswrapper[4702]: I1007 12:27:57.764282 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:27:57 crc kubenswrapper[4702]: I1007 12:27:57.765133 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:27:57 crc kubenswrapper[4702]: I1007 12:27:57.811998 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:27:57 crc kubenswrapper[4702]: I1007 12:27:57.855385 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dlh8f" podStartSLOduration=5.930111807 podStartE2EDuration="1m20.855353893s" podCreationTimestamp="2025-10-07 12:26:37 +0000 UTC" firstStartedPulling="2025-10-07 12:26:38.963240128 +0000 UTC m=+155.138976244" lastFinishedPulling="2025-10-07 12:27:53.888482214 +0000 UTC m=+230.064218330" observedRunningTime="2025-10-07 12:27:55.710744166 +0000 UTC m=+231.886480272" watchObservedRunningTime="2025-10-07 12:27:57.855353893 +0000 UTC m=+234.031090009" Oct 07 12:27:59 crc kubenswrapper[4702]: I1007 12:27:59.057589 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:27:59 crc kubenswrapper[4702]: I1007 12:27:59.117946 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.086271 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5mxz"] Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.088013 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f5mxz" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="registry-server" containerID="cri-o://8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f" gracePeriod=2 Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.500546 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.585976 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-utilities\") pod \"2247b061-f50a-40d4-bbab-b25a6284decf\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.586118 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-catalog-content\") pod \"2247b061-f50a-40d4-bbab-b25a6284decf\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.586174 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd2kf\" (UniqueName: \"kubernetes.io/projected/2247b061-f50a-40d4-bbab-b25a6284decf-kube-api-access-kd2kf\") pod \"2247b061-f50a-40d4-bbab-b25a6284decf\" (UID: \"2247b061-f50a-40d4-bbab-b25a6284decf\") " Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.587041 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-utilities" (OuterVolumeSpecName: "utilities") pod "2247b061-f50a-40d4-bbab-b25a6284decf" (UID: "2247b061-f50a-40d4-bbab-b25a6284decf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.595785 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2247b061-f50a-40d4-bbab-b25a6284decf-kube-api-access-kd2kf" (OuterVolumeSpecName: "kube-api-access-kd2kf") pod "2247b061-f50a-40d4-bbab-b25a6284decf" (UID: "2247b061-f50a-40d4-bbab-b25a6284decf"). InnerVolumeSpecName "kube-api-access-kd2kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.680309 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2247b061-f50a-40d4-bbab-b25a6284decf" (UID: "2247b061-f50a-40d4-bbab-b25a6284decf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.687652 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.687683 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b061-f50a-40d4-bbab-b25a6284decf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.687698 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd2kf\" (UniqueName: \"kubernetes.io/projected/2247b061-f50a-40d4-bbab-b25a6284decf-kube-api-access-kd2kf\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.741853 4702 generic.go:334] "Generic (PLEG): container finished" podID="2247b061-f50a-40d4-bbab-b25a6284decf" containerID="8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f" exitCode=0 Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.741959 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerDied","Data":"8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f"} Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.742038 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5mxz" event={"ID":"2247b061-f50a-40d4-bbab-b25a6284decf","Type":"ContainerDied","Data":"e6066b356a4ac2cae9d316c50f01a5ef62d5456d64a5cf4db27efb77f285a58d"} Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.742071 4702 scope.go:117] "RemoveContainer" containerID="8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.741984 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5mxz" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.766644 4702 scope.go:117] "RemoveContainer" containerID="4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.772104 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5mxz"] Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.781549 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f5mxz"] Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.806812 4702 scope.go:117] "RemoveContainer" containerID="986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.822423 4702 scope.go:117] "RemoveContainer" containerID="8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f" Oct 07 12:28:03 crc kubenswrapper[4702]: E1007 12:28:03.822874 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f\": container with ID starting with 8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f not found: ID does not exist" containerID="8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.822930 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f"} err="failed to get container status \"8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f\": rpc error: code = NotFound desc = could not find container \"8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f\": container with ID starting with 8270185af831cbf35b3c10c94a7379f1d771a09cb0d4dd63f009a028d8e4b06f not found: ID does not exist" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.822962 4702 scope.go:117] "RemoveContainer" containerID="4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc" Oct 07 12:28:03 crc kubenswrapper[4702]: E1007 12:28:03.823413 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc\": container with ID starting with 4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc not found: ID does not exist" containerID="4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.823434 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc"} err="failed to get container status \"4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc\": rpc error: code = NotFound desc = could not find container \"4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc\": container with ID starting with 4d42b98b289ec96df462d43daa38896aa1d8ee8a1c223a72257a57e94510e3fc not found: ID does not exist" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.823448 4702 scope.go:117] "RemoveContainer" containerID="986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd" Oct 07 12:28:03 crc kubenswrapper[4702]: E1007 12:28:03.823684 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd\": container with ID starting with 986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd not found: ID does not exist" containerID="986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd" Oct 07 12:28:03 crc kubenswrapper[4702]: I1007 12:28:03.823710 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd"} err="failed to get container status \"986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd\": rpc error: code = NotFound desc = could not find container \"986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd\": container with ID starting with 986aecb9b723e00841939724958c36d86a23c0fa666d27989a192003aa38fabd not found: ID does not exist" Oct 07 12:28:04 crc kubenswrapper[4702]: I1007 12:28:04.649611 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" path="/var/lib/kubelet/pods/2247b061-f50a-40d4-bbab-b25a6284decf/volumes" Oct 07 12:28:06 crc kubenswrapper[4702]: I1007 12:28:06.498000 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lbqfj"] Oct 07 12:28:07 crc kubenswrapper[4702]: I1007 12:28:07.805776 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:28:07 crc kubenswrapper[4702]: I1007 12:28:07.852365 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlh8f"] Oct 07 12:28:08 crc kubenswrapper[4702]: I1007 12:28:08.767802 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dlh8f" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="registry-server" containerID="cri-o://b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb" gracePeriod=2 Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.158344 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.266738 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-utilities\") pod \"23e3398a-ac68-4207-9475-3d072d1ce3e4\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.266817 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7t8c\" (UniqueName: \"kubernetes.io/projected/23e3398a-ac68-4207-9475-3d072d1ce3e4-kube-api-access-g7t8c\") pod \"23e3398a-ac68-4207-9475-3d072d1ce3e4\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.266950 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-catalog-content\") pod \"23e3398a-ac68-4207-9475-3d072d1ce3e4\" (UID: \"23e3398a-ac68-4207-9475-3d072d1ce3e4\") " Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.267756 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-utilities" (OuterVolumeSpecName: "utilities") pod "23e3398a-ac68-4207-9475-3d072d1ce3e4" (UID: "23e3398a-ac68-4207-9475-3d072d1ce3e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.273366 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23e3398a-ac68-4207-9475-3d072d1ce3e4-kube-api-access-g7t8c" (OuterVolumeSpecName: "kube-api-access-g7t8c") pod "23e3398a-ac68-4207-9475-3d072d1ce3e4" (UID: "23e3398a-ac68-4207-9475-3d072d1ce3e4"). InnerVolumeSpecName "kube-api-access-g7t8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.283312 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23e3398a-ac68-4207-9475-3d072d1ce3e4" (UID: "23e3398a-ac68-4207-9475-3d072d1ce3e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.368008 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.368066 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7t8c\" (UniqueName: \"kubernetes.io/projected/23e3398a-ac68-4207-9475-3d072d1ce3e4-kube-api-access-g7t8c\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.368079 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23e3398a-ac68-4207-9475-3d072d1ce3e4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.777351 4702 generic.go:334] "Generic (PLEG): container finished" podID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerID="b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb" exitCode=0 Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.777436 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerDied","Data":"b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb"} Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.777467 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlh8f" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.778124 4702 scope.go:117] "RemoveContainer" containerID="b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.778024 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlh8f" event={"ID":"23e3398a-ac68-4207-9475-3d072d1ce3e4","Type":"ContainerDied","Data":"9dd7924696842bf6dd552d4890c01cf7a626246f1e66c9484fe8a29f5c4c3aba"} Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.796887 4702 scope.go:117] "RemoveContainer" containerID="983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.821955 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlh8f"] Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.824631 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlh8f"] Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.832293 4702 scope.go:117] "RemoveContainer" containerID="e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.853195 4702 scope.go:117] "RemoveContainer" containerID="b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb" Oct 07 12:28:09 crc kubenswrapper[4702]: E1007 12:28:09.853916 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb\": container with ID starting with b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb not found: ID does not exist" containerID="b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.854050 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb"} err="failed to get container status \"b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb\": rpc error: code = NotFound desc = could not find container \"b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb\": container with ID starting with b09cff3da70046806182a08d40a2c895084aa47b07ebc879c4bf1ca9b38743bb not found: ID does not exist" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.854151 4702 scope.go:117] "RemoveContainer" containerID="983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad" Oct 07 12:28:09 crc kubenswrapper[4702]: E1007 12:28:09.854607 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad\": container with ID starting with 983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad not found: ID does not exist" containerID="983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.854696 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad"} err="failed to get container status \"983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad\": rpc error: code = NotFound desc = could not find container \"983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad\": container with ID starting with 983844c37b10070848f605e2b6c96954b691819fa682b2290270a7a5eef369ad not found: ID does not exist" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.854771 4702 scope.go:117] "RemoveContainer" containerID="e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c" Oct 07 12:28:09 crc kubenswrapper[4702]: E1007 12:28:09.855315 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c\": container with ID starting with e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c not found: ID does not exist" containerID="e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c" Oct 07 12:28:09 crc kubenswrapper[4702]: I1007 12:28:09.855409 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c"} err="failed to get container status \"e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c\": rpc error: code = NotFound desc = could not find container \"e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c\": container with ID starting with e9e6aefa0d2e6074495477467d44373bdfd32227b9689dd5a1957be37c98565c not found: ID does not exist" Oct 07 12:28:10 crc kubenswrapper[4702]: I1007 12:28:10.649356 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" path="/var/lib/kubelet/pods/23e3398a-ac68-4207-9475-3d072d1ce3e4/volumes" Oct 07 12:28:31 crc kubenswrapper[4702]: I1007 12:28:31.540273 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" podUID="364be77d-ab70-49ff-9fa4-2b42f49b718b" containerName="oauth-openshift" containerID="cri-o://f390836ac8a5aa23517284250ef14965f14ad56bf4e266fb32fb752154359275" gracePeriod=15 Oct 07 12:28:31 crc kubenswrapper[4702]: I1007 12:28:31.918878 4702 generic.go:334] "Generic (PLEG): container finished" podID="364be77d-ab70-49ff-9fa4-2b42f49b718b" containerID="f390836ac8a5aa23517284250ef14965f14ad56bf4e266fb32fb752154359275" exitCode=0 Oct 07 12:28:31 crc kubenswrapper[4702]: I1007 12:28:31.919121 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" event={"ID":"364be77d-ab70-49ff-9fa4-2b42f49b718b","Type":"ContainerDied","Data":"f390836ac8a5aa23517284250ef14965f14ad56bf4e266fb32fb752154359275"} Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.055343 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.092013 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-8kc7h"] Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.092614 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.092704 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.092828 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.092935 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.093116 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.093198 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.093284 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.093348 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.093432 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.093501 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.093634 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.093708 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.093780 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.093850 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.093966 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.094045 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.094125 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.094191 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.094248 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.094304 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="extract-utilities" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.094368 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed634baa-b36c-4a65-b8eb-84cf925a17f2" containerName="pruner" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.094424 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed634baa-b36c-4a65-b8eb-84cf925a17f2" containerName="pruner" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.094479 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364be77d-ab70-49ff-9fa4-2b42f49b718b" containerName="oauth-openshift" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.094534 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="364be77d-ab70-49ff-9fa4-2b42f49b718b" containerName="oauth-openshift" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.094592 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.094791 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="extract-content" Oct 07 12:28:32 crc kubenswrapper[4702]: E1007 12:28:32.094886 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095012 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095218 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e06f55-1237-41bb-bf4e-afe9a3342f6a" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095299 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e3398a-ac68-4207-9475-3d072d1ce3e4" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095358 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed634baa-b36c-4a65-b8eb-84cf925a17f2" containerName="pruner" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095413 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="27b2dc0e-3eab-47f5-a64d-b6b74f720dcd" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095479 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="364be77d-ab70-49ff-9fa4-2b42f49b718b" containerName="oauth-openshift" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.095541 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2247b061-f50a-40d4-bbab-b25a6284decf" containerName="registry-server" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.096163 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.118905 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-8kc7h"] Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.197851 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-cliconfig\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.198194 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-session\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.198332 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-service-ca\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.198494 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-login\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.198644 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-provider-selection\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.198783 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-router-certs\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.198884 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-error\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199007 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-idp-0-file-data\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199143 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-serving-cert\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199260 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn25n\" (UniqueName: \"kubernetes.io/projected/364be77d-ab70-49ff-9fa4-2b42f49b718b-kube-api-access-sn25n\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199372 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-trusted-ca-bundle\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199004 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199484 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199498 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-dir\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199700 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199809 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-policies\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199928 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-ocp-branding-template\") pod \"364be77d-ab70-49ff-9fa4-2b42f49b718b\" (UID: \"364be77d-ab70-49ff-9fa4-2b42f49b718b\") " Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.199951 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200016 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200048 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200075 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6cf697d7-d314-43bf-bb86-2fc76b644bef-audit-dir\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200097 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200127 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200169 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200187 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200205 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200280 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200303 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200345 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200416 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtzbc\" (UniqueName: \"kubernetes.io/projected/6cf697d7-d314-43bf-bb86-2fc76b644bef-kube-api-access-mtzbc\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200447 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-audit-policies\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200470 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200489 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200529 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200544 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200557 4702 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200569 4702 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.200582 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.205262 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.218126 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.218263 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.219112 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364be77d-ab70-49ff-9fa4-2b42f49b718b-kube-api-access-sn25n" (OuterVolumeSpecName: "kube-api-access-sn25n") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "kube-api-access-sn25n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.219740 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.220097 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.220164 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.220543 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.219680 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "364be77d-ab70-49ff-9fa4-2b42f49b718b" (UID: "364be77d-ab70-49ff-9fa4-2b42f49b718b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.301594 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.301986 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtzbc\" (UniqueName: \"kubernetes.io/projected/6cf697d7-d314-43bf-bb86-2fc76b644bef-kube-api-access-mtzbc\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.302544 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-audit-policies\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.303372 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.304223 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.303327 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-audit-policies\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.304397 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.304655 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.304787 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6cf697d7-d314-43bf-bb86-2fc76b644bef-audit-dir\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.304879 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.304977 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6cf697d7-d314-43bf-bb86-2fc76b644bef-audit-dir\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.305084 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.305200 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.305320 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.305860 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.306716 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.306820 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.306987 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.307220 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.307403 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.308167 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.308382 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.307697 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.308844 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.308950 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309033 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309127 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309188 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309285 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309370 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309456 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309547 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/364be77d-ab70-49ff-9fa4-2b42f49b718b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.309634 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn25n\" (UniqueName: \"kubernetes.io/projected/364be77d-ab70-49ff-9fa4-2b42f49b718b-kube-api-access-sn25n\") on node \"crc\" DevicePath \"\"" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.311433 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.313126 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.314360 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6cf697d7-d314-43bf-bb86-2fc76b644bef-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.319044 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtzbc\" (UniqueName: \"kubernetes.io/projected/6cf697d7-d314-43bf-bb86-2fc76b644bef-kube-api-access-mtzbc\") pod \"oauth-openshift-594b4c596d-8kc7h\" (UID: \"6cf697d7-d314-43bf-bb86-2fc76b644bef\") " pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.419863 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.657778 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-8kc7h"] Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.926725 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" event={"ID":"6cf697d7-d314-43bf-bb86-2fc76b644bef","Type":"ContainerStarted","Data":"3ab2b4e9a2f85a37c75f94aac764021454f2e6466fc033ee1fee81e6d54eae97"} Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.927301 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" event={"ID":"6cf697d7-d314-43bf-bb86-2fc76b644bef","Type":"ContainerStarted","Data":"c39f0dfc2cac746d0dc49e37d966526469f1e4a4a3979f707057eee443847f87"} Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.927330 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.929003 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" event={"ID":"364be77d-ab70-49ff-9fa4-2b42f49b718b","Type":"ContainerDied","Data":"b5e26f68a13c510b3ce746633acc427a168512fc93679a2edd036d5330e572b9"} Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.929058 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lbqfj" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.929107 4702 scope.go:117] "RemoveContainer" containerID="f390836ac8a5aa23517284250ef14965f14ad56bf4e266fb32fb752154359275" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.929032 4702 patch_prober.go:28] interesting pod/oauth-openshift-594b4c596d-8kc7h container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.54:6443/healthz\": dial tcp 10.217.0.54:6443: connect: connection refused" start-of-body= Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.929252 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" podUID="6cf697d7-d314-43bf-bb86-2fc76b644bef" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.54:6443/healthz\": dial tcp 10.217.0.54:6443: connect: connection refused" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.950591 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" podStartSLOduration=26.950568909 podStartE2EDuration="26.950568909s" podCreationTimestamp="2025-10-07 12:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:28:32.950355762 +0000 UTC m=+269.126091878" watchObservedRunningTime="2025-10-07 12:28:32.950568909 +0000 UTC m=+269.126305035" Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.967771 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lbqfj"] Oct 07 12:28:32 crc kubenswrapper[4702]: I1007 12:28:32.970777 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lbqfj"] Oct 07 12:28:33 crc kubenswrapper[4702]: I1007 12:28:33.941030 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-594b4c596d-8kc7h" Oct 07 12:28:34 crc kubenswrapper[4702]: I1007 12:28:34.657816 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364be77d-ab70-49ff-9fa4-2b42f49b718b" path="/var/lib/kubelet/pods/364be77d-ab70-49ff-9fa4-2b42f49b718b/volumes" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.601540 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlzrn"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.603235 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zlzrn" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="registry-server" containerID="cri-o://e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d" gracePeriod=30 Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.625202 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bhwkv"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.626231 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bhwkv" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="registry-server" containerID="cri-o://5dbbbb8db5d824382b20725348694e995a54c726122b01ca035ce53ca179cd12" gracePeriod=30 Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.638733 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcw2q"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.639147 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" containerID="cri-o://e384fa09c91796f1e92901132c8b696d6d9e0947e072110f88ac54ffbf2a36e9" gracePeriod=30 Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.644368 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xtg2"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.644775 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9xtg2" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="registry-server" containerID="cri-o://ca1d8e878af469855bd808fc19ba8d9c1342a05e288cd0bcf81a8de99ac33070" gracePeriod=30 Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.647066 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zjkc"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.647322 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8zjkc" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="registry-server" containerID="cri-o://0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3" gracePeriod=30 Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.659937 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8r8v4"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.661323 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.667811 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8r8v4"] Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.739813 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/76e55658-e2c9-4e68-b04a-a8ebd620e355-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.739863 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/76e55658-e2c9-4e68-b04a-a8ebd620e355-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.739987 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h52j\" (UniqueName: \"kubernetes.io/projected/76e55658-e2c9-4e68-b04a-a8ebd620e355-kube-api-access-6h52j\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.841039 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/76e55658-e2c9-4e68-b04a-a8ebd620e355-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.841096 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/76e55658-e2c9-4e68-b04a-a8ebd620e355-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.841141 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h52j\" (UniqueName: \"kubernetes.io/projected/76e55658-e2c9-4e68-b04a-a8ebd620e355-kube-api-access-6h52j\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.842884 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/76e55658-e2c9-4e68-b04a-a8ebd620e355-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.849233 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/76e55658-e2c9-4e68-b04a-a8ebd620e355-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.862645 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h52j\" (UniqueName: \"kubernetes.io/projected/76e55658-e2c9-4e68-b04a-a8ebd620e355-kube-api-access-6h52j\") pod \"marketplace-operator-79b997595-8r8v4\" (UID: \"76e55658-e2c9-4e68-b04a-a8ebd620e355\") " pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:01 crc kubenswrapper[4702]: I1007 12:29:01.985911 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.082670 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.084381 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.139666 4702 generic.go:334] "Generic (PLEG): container finished" podID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerID="e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d" exitCode=0 Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.139771 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerDied","Data":"e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.139817 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zlzrn" event={"ID":"536f8e03-4e07-489d-8505-a6fc83e98dc2","Type":"ContainerDied","Data":"09a6cb6c95a08f8533a3fa18301877e6483ed60900fe2beb458e03b941e1a507"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.139837 4702 scope.go:117] "RemoveContainer" containerID="e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.140065 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zlzrn" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.141098 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.146478 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb9hf\" (UniqueName: \"kubernetes.io/projected/988399d3-c52f-450e-a74f-3fd6bbf03c1f-kube-api-access-cb9hf\") pod \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.146583 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-utilities\") pod \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.146616 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-catalog-content\") pod \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\" (UID: \"988399d3-c52f-450e-a74f-3fd6bbf03c1f\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.146656 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww5dc\" (UniqueName: \"kubernetes.io/projected/536f8e03-4e07-489d-8505-a6fc83e98dc2-kube-api-access-ww5dc\") pod \"536f8e03-4e07-489d-8505-a6fc83e98dc2\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.146709 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-utilities\") pod \"536f8e03-4e07-489d-8505-a6fc83e98dc2\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.146755 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-catalog-content\") pod \"536f8e03-4e07-489d-8505-a6fc83e98dc2\" (UID: \"536f8e03-4e07-489d-8505-a6fc83e98dc2\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.148875 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-utilities" (OuterVolumeSpecName: "utilities") pod "536f8e03-4e07-489d-8505-a6fc83e98dc2" (UID: "536f8e03-4e07-489d-8505-a6fc83e98dc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.149378 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-utilities" (OuterVolumeSpecName: "utilities") pod "988399d3-c52f-450e-a74f-3fd6bbf03c1f" (UID: "988399d3-c52f-450e-a74f-3fd6bbf03c1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.153407 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.154006 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/988399d3-c52f-450e-a74f-3fd6bbf03c1f-kube-api-access-cb9hf" (OuterVolumeSpecName: "kube-api-access-cb9hf") pod "988399d3-c52f-450e-a74f-3fd6bbf03c1f" (UID: "988399d3-c52f-450e-a74f-3fd6bbf03c1f"). InnerVolumeSpecName "kube-api-access-cb9hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.154596 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.157406 4702 generic.go:334] "Generic (PLEG): container finished" podID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerID="0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3" exitCode=0 Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.157480 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjkc" event={"ID":"988399d3-c52f-450e-a74f-3fd6bbf03c1f","Type":"ContainerDied","Data":"0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.157515 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zjkc" event={"ID":"988399d3-c52f-450e-a74f-3fd6bbf03c1f","Type":"ContainerDied","Data":"3ce2e0705658d2f67cb6d23c85cdfd07ba6f4e3c231250afa29be001692ff98c"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.157573 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zjkc" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.159385 4702 generic.go:334] "Generic (PLEG): container finished" podID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerID="e384fa09c91796f1e92901132c8b696d6d9e0947e072110f88ac54ffbf2a36e9" exitCode=0 Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.159457 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" event={"ID":"6f164dd1-c728-4d35-b7f8-12fca172497d","Type":"ContainerDied","Data":"e384fa09c91796f1e92901132c8b696d6d9e0947e072110f88ac54ffbf2a36e9"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.159514 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcw2q" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.159617 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536f8e03-4e07-489d-8505-a6fc83e98dc2-kube-api-access-ww5dc" (OuterVolumeSpecName: "kube-api-access-ww5dc") pod "536f8e03-4e07-489d-8505-a6fc83e98dc2" (UID: "536f8e03-4e07-489d-8505-a6fc83e98dc2"). InnerVolumeSpecName "kube-api-access-ww5dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.165657 4702 generic.go:334] "Generic (PLEG): container finished" podID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerID="ca1d8e878af469855bd808fc19ba8d9c1342a05e288cd0bcf81a8de99ac33070" exitCode=0 Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.165739 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xtg2" event={"ID":"cf22b7ba-6aac-40b0-b25e-72e092658ac6","Type":"ContainerDied","Data":"ca1d8e878af469855bd808fc19ba8d9c1342a05e288cd0bcf81a8de99ac33070"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.165845 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xtg2" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.186840 4702 scope.go:117] "RemoveContainer" containerID="5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.207202 4702 generic.go:334] "Generic (PLEG): container finished" podID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerID="5dbbbb8db5d824382b20725348694e995a54c726122b01ca035ce53ca179cd12" exitCode=0 Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.207269 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhwkv" event={"ID":"55a80c96-1beb-4bc7-b84a-536e43e25b82","Type":"ContainerDied","Data":"5dbbbb8db5d824382b20725348694e995a54c726122b01ca035ce53ca179cd12"} Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.207414 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhwkv" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.229131 4702 scope.go:117] "RemoveContainer" containerID="4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.239383 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "536f8e03-4e07-489d-8505-a6fc83e98dc2" (UID: "536f8e03-4e07-489d-8505-a6fc83e98dc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248473 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-catalog-content\") pod \"55a80c96-1beb-4bc7-b84a-536e43e25b82\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248537 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-trusted-ca\") pod \"6f164dd1-c728-4d35-b7f8-12fca172497d\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248566 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-catalog-content\") pod \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248645 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-utilities\") pod \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248666 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw5hl\" (UniqueName: \"kubernetes.io/projected/55a80c96-1beb-4bc7-b84a-536e43e25b82-kube-api-access-sw5hl\") pod \"55a80c96-1beb-4bc7-b84a-536e43e25b82\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248686 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-operator-metrics\") pod \"6f164dd1-c728-4d35-b7f8-12fca172497d\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248715 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f49sq\" (UniqueName: \"kubernetes.io/projected/cf22b7ba-6aac-40b0-b25e-72e092658ac6-kube-api-access-f49sq\") pod \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\" (UID: \"cf22b7ba-6aac-40b0-b25e-72e092658ac6\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248771 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jngnc\" (UniqueName: \"kubernetes.io/projected/6f164dd1-c728-4d35-b7f8-12fca172497d-kube-api-access-jngnc\") pod \"6f164dd1-c728-4d35-b7f8-12fca172497d\" (UID: \"6f164dd1-c728-4d35-b7f8-12fca172497d\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.248886 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-utilities\") pod \"55a80c96-1beb-4bc7-b84a-536e43e25b82\" (UID: \"55a80c96-1beb-4bc7-b84a-536e43e25b82\") " Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.249102 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.249121 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww5dc\" (UniqueName: \"kubernetes.io/projected/536f8e03-4e07-489d-8505-a6fc83e98dc2-kube-api-access-ww5dc\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.249135 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.249143 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/536f8e03-4e07-489d-8505-a6fc83e98dc2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.249152 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb9hf\" (UniqueName: \"kubernetes.io/projected/988399d3-c52f-450e-a74f-3fd6bbf03c1f-kube-api-access-cb9hf\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.249834 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-utilities" (OuterVolumeSpecName: "utilities") pod "55a80c96-1beb-4bc7-b84a-536e43e25b82" (UID: "55a80c96-1beb-4bc7-b84a-536e43e25b82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.252926 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a80c96-1beb-4bc7-b84a-536e43e25b82-kube-api-access-sw5hl" (OuterVolumeSpecName: "kube-api-access-sw5hl") pod "55a80c96-1beb-4bc7-b84a-536e43e25b82" (UID: "55a80c96-1beb-4bc7-b84a-536e43e25b82"). InnerVolumeSpecName "kube-api-access-sw5hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.253598 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f164dd1-c728-4d35-b7f8-12fca172497d-kube-api-access-jngnc" (OuterVolumeSpecName: "kube-api-access-jngnc") pod "6f164dd1-c728-4d35-b7f8-12fca172497d" (UID: "6f164dd1-c728-4d35-b7f8-12fca172497d"). InnerVolumeSpecName "kube-api-access-jngnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.254369 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-utilities" (OuterVolumeSpecName: "utilities") pod "cf22b7ba-6aac-40b0-b25e-72e092658ac6" (UID: "cf22b7ba-6aac-40b0-b25e-72e092658ac6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.256422 4702 scope.go:117] "RemoveContainer" containerID="e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.257431 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6f164dd1-c728-4d35-b7f8-12fca172497d" (UID: "6f164dd1-c728-4d35-b7f8-12fca172497d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: E1007 12:29:02.257761 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d\": container with ID starting with e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d not found: ID does not exist" containerID="e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.257813 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d"} err="failed to get container status \"e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d\": rpc error: code = NotFound desc = could not find container \"e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d\": container with ID starting with e518f8098aad23818b9c2eff8976d55a0cd3e9a05d75a731515b57bbdc529d3d not found: ID does not exist" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.257858 4702 scope.go:117] "RemoveContainer" containerID="5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4" Oct 07 12:29:02 crc kubenswrapper[4702]: E1007 12:29:02.258224 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4\": container with ID starting with 5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4 not found: ID does not exist" containerID="5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.258240 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4"} err="failed to get container status \"5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4\": rpc error: code = NotFound desc = could not find container \"5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4\": container with ID starting with 5bc0ef29b420b21be4933c09038a5ef335b2250d215a57860ff93c1bccef10d4 not found: ID does not exist" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.258253 4702 scope.go:117] "RemoveContainer" containerID="4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.258280 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6f164dd1-c728-4d35-b7f8-12fca172497d" (UID: "6f164dd1-c728-4d35-b7f8-12fca172497d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: E1007 12:29:02.259056 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297\": container with ID starting with 4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297 not found: ID does not exist" containerID="4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.259094 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297"} err="failed to get container status \"4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297\": rpc error: code = NotFound desc = could not find container \"4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297\": container with ID starting with 4a1185b2fca329e1158c50aeda99ffd84c88c5c32e0fbb0283fd18b344b9e297 not found: ID does not exist" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.259147 4702 scope.go:117] "RemoveContainer" containerID="0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.263825 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf22b7ba-6aac-40b0-b25e-72e092658ac6" (UID: "cf22b7ba-6aac-40b0-b25e-72e092658ac6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.268411 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf22b7ba-6aac-40b0-b25e-72e092658ac6-kube-api-access-f49sq" (OuterVolumeSpecName: "kube-api-access-f49sq") pod "cf22b7ba-6aac-40b0-b25e-72e092658ac6" (UID: "cf22b7ba-6aac-40b0-b25e-72e092658ac6"). InnerVolumeSpecName "kube-api-access-f49sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.289994 4702 scope.go:117] "RemoveContainer" containerID="99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.295069 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "988399d3-c52f-450e-a74f-3fd6bbf03c1f" (UID: "988399d3-c52f-450e-a74f-3fd6bbf03c1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.297188 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55a80c96-1beb-4bc7-b84a-536e43e25b82" (UID: "55a80c96-1beb-4bc7-b84a-536e43e25b82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.305725 4702 scope.go:117] "RemoveContainer" containerID="98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.319927 4702 scope.go:117] "RemoveContainer" containerID="0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3" Oct 07 12:29:02 crc kubenswrapper[4702]: E1007 12:29:02.320487 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3\": container with ID starting with 0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3 not found: ID does not exist" containerID="0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.320524 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3"} err="failed to get container status \"0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3\": rpc error: code = NotFound desc = could not find container \"0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3\": container with ID starting with 0b14fb49d3ee5d969ec451e95144ffc61e50c54dfaee29420eb72c892f77eae3 not found: ID does not exist" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.320559 4702 scope.go:117] "RemoveContainer" containerID="99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426" Oct 07 12:29:02 crc kubenswrapper[4702]: E1007 12:29:02.321012 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426\": container with ID starting with 99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426 not found: ID does not exist" containerID="99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.321073 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426"} err="failed to get container status \"99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426\": rpc error: code = NotFound desc = could not find container \"99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426\": container with ID starting with 99d5951f7f98c913dff72147ca895c92d3794804e02441efba9bf5b68a122426 not found: ID does not exist" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.321121 4702 scope.go:117] "RemoveContainer" containerID="98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f" Oct 07 12:29:02 crc kubenswrapper[4702]: E1007 12:29:02.322172 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f\": container with ID starting with 98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f not found: ID does not exist" containerID="98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.322203 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f"} err="failed to get container status \"98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f\": rpc error: code = NotFound desc = could not find container \"98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f\": container with ID starting with 98754af6d1a832f8804f477c573f58820ff274913b71025817e5c7ab6a1b2c3f not found: ID does not exist" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.322219 4702 scope.go:117] "RemoveContainer" containerID="e384fa09c91796f1e92901132c8b696d6d9e0947e072110f88ac54ffbf2a36e9" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.335726 4702 scope.go:117] "RemoveContainer" containerID="ca1d8e878af469855bd808fc19ba8d9c1342a05e288cd0bcf81a8de99ac33070" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.347044 4702 scope.go:117] "RemoveContainer" containerID="f34ea6b84816395c80aff712f49e84ab6bb5466373cbbd3835a1cd2abab8a1f4" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350777 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350802 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350814 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350824 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988399d3-c52f-450e-a74f-3fd6bbf03c1f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350835 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf22b7ba-6aac-40b0-b25e-72e092658ac6-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350877 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw5hl\" (UniqueName: \"kubernetes.io/projected/55a80c96-1beb-4bc7-b84a-536e43e25b82-kube-api-access-sw5hl\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350887 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f164dd1-c728-4d35-b7f8-12fca172497d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350912 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f49sq\" (UniqueName: \"kubernetes.io/projected/cf22b7ba-6aac-40b0-b25e-72e092658ac6-kube-api-access-f49sq\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350925 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jngnc\" (UniqueName: \"kubernetes.io/projected/6f164dd1-c728-4d35-b7f8-12fca172497d-kube-api-access-jngnc\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.350934 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a80c96-1beb-4bc7-b84a-536e43e25b82-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.358597 4702 scope.go:117] "RemoveContainer" containerID="6eb50bdd69ea6a3417672c0f997ca03d566460d0ad375d18f641d95ca21857f2" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.370288 4702 scope.go:117] "RemoveContainer" containerID="5dbbbb8db5d824382b20725348694e995a54c726122b01ca035ce53ca179cd12" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.383868 4702 scope.go:117] "RemoveContainer" containerID="4a82f11900f5e9f77cb3cb83bf699e4f436a94e6da066831eab0cf924ce85036" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.397463 4702 scope.go:117] "RemoveContainer" containerID="d751b3b289a2172e7c7d416b32357119760e87748a9e0381a731f38fa491ca8c" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.474457 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zlzrn"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.482328 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zlzrn"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.490787 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8r8v4"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.494277 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcw2q"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.506425 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcw2q"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.520234 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xtg2"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.524085 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xtg2"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.533351 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zjkc"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.537607 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8zjkc"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.547728 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bhwkv"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.550204 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bhwkv"] Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.649268 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" path="/var/lib/kubelet/pods/536f8e03-4e07-489d-8505-a6fc83e98dc2/volumes" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.650934 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" path="/var/lib/kubelet/pods/55a80c96-1beb-4bc7-b84a-536e43e25b82/volumes" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.651903 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" path="/var/lib/kubelet/pods/6f164dd1-c728-4d35-b7f8-12fca172497d/volumes" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.652424 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" path="/var/lib/kubelet/pods/988399d3-c52f-450e-a74f-3fd6bbf03c1f/volumes" Oct 07 12:29:02 crc kubenswrapper[4702]: I1007 12:29:02.652999 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" path="/var/lib/kubelet/pods/cf22b7ba-6aac-40b0-b25e-72e092658ac6/volumes" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.214103 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" event={"ID":"76e55658-e2c9-4e68-b04a-a8ebd620e355","Type":"ContainerStarted","Data":"34fa507ba43071407c4dc20d39cb0bab7c55db8945723ec2abb3dc25a38704a8"} Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.214666 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" event={"ID":"76e55658-e2c9-4e68-b04a-a8ebd620e355","Type":"ContainerStarted","Data":"67cf729c20280bf4c37962eca96cb3246cb303781547fb0e7ac2111500038065"} Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.214695 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.220230 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.250523 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8r8v4" podStartSLOduration=2.250494812 podStartE2EDuration="2.250494812s" podCreationTimestamp="2025-10-07 12:29:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:29:03.231259383 +0000 UTC m=+299.406995509" watchObservedRunningTime="2025-10-07 12:29:03.250494812 +0000 UTC m=+299.426230928" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.827663 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m4xtt"] Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.827935 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.827957 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.827977 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.827985 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828001 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828013 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828028 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828038 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828051 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828059 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828069 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828078 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828087 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828095 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828104 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828111 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828120 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828127 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828137 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828144 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828156 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828163 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="extract-utilities" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828175 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828183 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: E1007 12:29:03.828193 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828693 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="extract-content" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828810 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a80c96-1beb-4bc7-b84a-536e43e25b82" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828827 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="988399d3-c52f-450e-a74f-3fd6bbf03c1f" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828837 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="536f8e03-4e07-489d-8505-a6fc83e98dc2" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828844 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f164dd1-c728-4d35-b7f8-12fca172497d" containerName="marketplace-operator" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.828857 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf22b7ba-6aac-40b0-b25e-72e092658ac6" containerName="registry-server" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.829658 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.836065 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.845443 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4xtt"] Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.872813 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-utilities\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.872928 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf58v\" (UniqueName: \"kubernetes.io/projected/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-kube-api-access-bf58v\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.873085 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-catalog-content\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.974571 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-utilities\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.974667 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf58v\" (UniqueName: \"kubernetes.io/projected/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-kube-api-access-bf58v\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.974694 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-catalog-content\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.975195 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-utilities\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.975254 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-catalog-content\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:03 crc kubenswrapper[4702]: I1007 12:29:03.996984 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf58v\" (UniqueName: \"kubernetes.io/projected/bc501a5c-2f7a-4668-a581-6d8b2d0191e8-kube-api-access-bf58v\") pod \"redhat-marketplace-m4xtt\" (UID: \"bc501a5c-2f7a-4668-a581-6d8b2d0191e8\") " pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.025675 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-82sgm"] Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.026969 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.030628 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.041913 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82sgm"] Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.075586 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-utilities\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.075653 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6jzc\" (UniqueName: \"kubernetes.io/projected/35e85fb0-1e14-4990-a158-3f693778ddbe-kube-api-access-g6jzc\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.075717 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-catalog-content\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.147661 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.176696 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-utilities\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.176752 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6jzc\" (UniqueName: \"kubernetes.io/projected/35e85fb0-1e14-4990-a158-3f693778ddbe-kube-api-access-g6jzc\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.176792 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-catalog-content\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.177349 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-catalog-content\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.177619 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-utilities\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.197725 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6jzc\" (UniqueName: \"kubernetes.io/projected/35e85fb0-1e14-4990-a158-3f693778ddbe-kube-api-access-g6jzc\") pod \"redhat-operators-82sgm\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.344619 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4xtt"] Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.354173 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:04 crc kubenswrapper[4702]: I1007 12:29:04.554494 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82sgm"] Oct 07 12:29:05 crc kubenswrapper[4702]: I1007 12:29:05.235882 4702 generic.go:334] "Generic (PLEG): container finished" podID="bc501a5c-2f7a-4668-a581-6d8b2d0191e8" containerID="ae054a7d313af91e270620abad4c7a4a580703c2526bee1f1e3a1028da483615" exitCode=0 Oct 07 12:29:05 crc kubenswrapper[4702]: I1007 12:29:05.235939 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4xtt" event={"ID":"bc501a5c-2f7a-4668-a581-6d8b2d0191e8","Type":"ContainerDied","Data":"ae054a7d313af91e270620abad4c7a4a580703c2526bee1f1e3a1028da483615"} Oct 07 12:29:05 crc kubenswrapper[4702]: I1007 12:29:05.236512 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4xtt" event={"ID":"bc501a5c-2f7a-4668-a581-6d8b2d0191e8","Type":"ContainerStarted","Data":"6b0f4eb9bdec41e9ef9eef107f88f62754895392217799176e6d71da7c21aaaf"} Oct 07 12:29:05 crc kubenswrapper[4702]: I1007 12:29:05.238764 4702 generic.go:334] "Generic (PLEG): container finished" podID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerID="dcbdb1fb358e0f48df841e674187ed5490bb2147c13e4396ed6fb70618ff36bc" exitCode=0 Oct 07 12:29:05 crc kubenswrapper[4702]: I1007 12:29:05.238858 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerDied","Data":"dcbdb1fb358e0f48df841e674187ed5490bb2147c13e4396ed6fb70618ff36bc"} Oct 07 12:29:05 crc kubenswrapper[4702]: I1007 12:29:05.238919 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerStarted","Data":"49c1ee46f9c3c6fd1dd80fe6ce318796bd6fc35be3562a29d85188259767800e"} Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.240213 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-86dwm"] Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.244678 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.247962 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.248202 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-86dwm"] Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.269852 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4xtt" event={"ID":"bc501a5c-2f7a-4668-a581-6d8b2d0191e8","Type":"ContainerStarted","Data":"ad7aed34f9a0a8a7b0d98c5ae89a86fe2c33fb1aecb2260964d61a2f8a8d2ab2"} Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.272981 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerStarted","Data":"1a7c044a1765b68b759f88538d04c91631f1c80e7cedc24521c2bf257004e8bf"} Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.305821 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3119f5-1f29-4f75-ad92-ace559a6b148-catalog-content\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.305929 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3119f5-1f29-4f75-ad92-ace559a6b148-utilities\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.306003 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjj9l\" (UniqueName: \"kubernetes.io/projected/8b3119f5-1f29-4f75-ad92-ace559a6b148-kube-api-access-wjj9l\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.408188 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3119f5-1f29-4f75-ad92-ace559a6b148-catalog-content\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.408579 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3119f5-1f29-4f75-ad92-ace559a6b148-utilities\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.408793 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjj9l\" (UniqueName: \"kubernetes.io/projected/8b3119f5-1f29-4f75-ad92-ace559a6b148-kube-api-access-wjj9l\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.409163 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3119f5-1f29-4f75-ad92-ace559a6b148-utilities\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.417311 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3119f5-1f29-4f75-ad92-ace559a6b148-catalog-content\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.432361 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2ntvb"] Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.437718 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.442164 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.444997 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2ntvb"] Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.448594 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjj9l\" (UniqueName: \"kubernetes.io/projected/8b3119f5-1f29-4f75-ad92-ace559a6b148-kube-api-access-wjj9l\") pod \"certified-operators-86dwm\" (UID: \"8b3119f5-1f29-4f75-ad92-ace559a6b148\") " pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.509842 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sqcs\" (UniqueName: \"kubernetes.io/projected/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-kube-api-access-6sqcs\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.509919 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-catalog-content\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.510294 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-utilities\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.611523 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sqcs\" (UniqueName: \"kubernetes.io/projected/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-kube-api-access-6sqcs\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.611574 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-catalog-content\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.611605 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-utilities\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.612601 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-catalog-content\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.612662 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-utilities\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.629110 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sqcs\" (UniqueName: \"kubernetes.io/projected/de46fdcb-23b3-4c75-baf4-584a8b29d3d3-kube-api-access-6sqcs\") pod \"community-operators-2ntvb\" (UID: \"de46fdcb-23b3-4c75-baf4-584a8b29d3d3\") " pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.684023 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.782311 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:06 crc kubenswrapper[4702]: I1007 12:29:06.990711 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-86dwm"] Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.249088 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2ntvb"] Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.278639 4702 generic.go:334] "Generic (PLEG): container finished" podID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerID="1a7c044a1765b68b759f88538d04c91631f1c80e7cedc24521c2bf257004e8bf" exitCode=0 Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.278700 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerDied","Data":"1a7c044a1765b68b759f88538d04c91631f1c80e7cedc24521c2bf257004e8bf"} Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.292292 4702 generic.go:334] "Generic (PLEG): container finished" podID="8b3119f5-1f29-4f75-ad92-ace559a6b148" containerID="d316b86294e5fdeecea7ee25f0ff961574369418ad9b1d27af54091eda324898" exitCode=0 Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.292372 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-86dwm" event={"ID":"8b3119f5-1f29-4f75-ad92-ace559a6b148","Type":"ContainerDied","Data":"d316b86294e5fdeecea7ee25f0ff961574369418ad9b1d27af54091eda324898"} Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.292409 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-86dwm" event={"ID":"8b3119f5-1f29-4f75-ad92-ace559a6b148","Type":"ContainerStarted","Data":"0760c2bf3ab8c34e305cc1fb651a916b20b3eb0c71ccde4ed98055826a0f3f63"} Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.295821 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ntvb" event={"ID":"de46fdcb-23b3-4c75-baf4-584a8b29d3d3","Type":"ContainerStarted","Data":"ea985fa5ad379821b104fa22972abcc756d0a404d3f82aa9cb365ef0ac4b6067"} Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.300189 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4xtt" event={"ID":"bc501a5c-2f7a-4668-a581-6d8b2d0191e8","Type":"ContainerDied","Data":"ad7aed34f9a0a8a7b0d98c5ae89a86fe2c33fb1aecb2260964d61a2f8a8d2ab2"} Oct 07 12:29:07 crc kubenswrapper[4702]: I1007 12:29:07.300050 4702 generic.go:334] "Generic (PLEG): container finished" podID="bc501a5c-2f7a-4668-a581-6d8b2d0191e8" containerID="ad7aed34f9a0a8a7b0d98c5ae89a86fe2c33fb1aecb2260964d61a2f8a8d2ab2" exitCode=0 Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.310129 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4xtt" event={"ID":"bc501a5c-2f7a-4668-a581-6d8b2d0191e8","Type":"ContainerStarted","Data":"4a37cf28df837d83e6a30652d3f1d1a0f1788fdf4683d0b7772bcebed82a6efd"} Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.314424 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerStarted","Data":"d67e9e4b2485b0d9807c469ef1e7b0a7018e4be99e90350410c7af0e24b4af31"} Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.316812 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-86dwm" event={"ID":"8b3119f5-1f29-4f75-ad92-ace559a6b148","Type":"ContainerStarted","Data":"5836c653b92c0bdad3d5559da8e8667a8511ee46382f90954dc651308e3c79ed"} Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.318368 4702 generic.go:334] "Generic (PLEG): container finished" podID="de46fdcb-23b3-4c75-baf4-584a8b29d3d3" containerID="8fe418015f951c552005f16c8c348cb6c3b7661adf7b1a18ddcc292ee33c5251" exitCode=0 Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.318408 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ntvb" event={"ID":"de46fdcb-23b3-4c75-baf4-584a8b29d3d3","Type":"ContainerDied","Data":"8fe418015f951c552005f16c8c348cb6c3b7661adf7b1a18ddcc292ee33c5251"} Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.337657 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m4xtt" podStartSLOduration=2.588825232 podStartE2EDuration="5.337627384s" podCreationTimestamp="2025-10-07 12:29:03 +0000 UTC" firstStartedPulling="2025-10-07 12:29:05.237264306 +0000 UTC m=+301.413000422" lastFinishedPulling="2025-10-07 12:29:07.986066418 +0000 UTC m=+304.161802574" observedRunningTime="2025-10-07 12:29:08.334945763 +0000 UTC m=+304.510681889" watchObservedRunningTime="2025-10-07 12:29:08.337627384 +0000 UTC m=+304.513363490" Oct 07 12:29:08 crc kubenswrapper[4702]: I1007 12:29:08.405655 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-82sgm" podStartSLOduration=1.795275868 podStartE2EDuration="4.405628451s" podCreationTimestamp="2025-10-07 12:29:04 +0000 UTC" firstStartedPulling="2025-10-07 12:29:05.240087771 +0000 UTC m=+301.415823897" lastFinishedPulling="2025-10-07 12:29:07.850440364 +0000 UTC m=+304.026176480" observedRunningTime="2025-10-07 12:29:08.377477474 +0000 UTC m=+304.553213640" watchObservedRunningTime="2025-10-07 12:29:08.405628451 +0000 UTC m=+304.581364567" Oct 07 12:29:09 crc kubenswrapper[4702]: I1007 12:29:09.326958 4702 generic.go:334] "Generic (PLEG): container finished" podID="8b3119f5-1f29-4f75-ad92-ace559a6b148" containerID="5836c653b92c0bdad3d5559da8e8667a8511ee46382f90954dc651308e3c79ed" exitCode=0 Oct 07 12:29:09 crc kubenswrapper[4702]: I1007 12:29:09.328730 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-86dwm" event={"ID":"8b3119f5-1f29-4f75-ad92-ace559a6b148","Type":"ContainerDied","Data":"5836c653b92c0bdad3d5559da8e8667a8511ee46382f90954dc651308e3c79ed"} Oct 07 12:29:10 crc kubenswrapper[4702]: I1007 12:29:10.335285 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-86dwm" event={"ID":"8b3119f5-1f29-4f75-ad92-ace559a6b148","Type":"ContainerStarted","Data":"b2966cdb1bae2fe0b596fef4b828152b8aebd0dbaa5cac8bb6adaee6c413b0f8"} Oct 07 12:29:10 crc kubenswrapper[4702]: I1007 12:29:10.336878 4702 generic.go:334] "Generic (PLEG): container finished" podID="de46fdcb-23b3-4c75-baf4-584a8b29d3d3" containerID="738a982880237c161bcb4011c72588fabecffca40735d1e5e34652b71f2c5c1a" exitCode=0 Oct 07 12:29:10 crc kubenswrapper[4702]: I1007 12:29:10.336987 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ntvb" event={"ID":"de46fdcb-23b3-4c75-baf4-584a8b29d3d3","Type":"ContainerDied","Data":"738a982880237c161bcb4011c72588fabecffca40735d1e5e34652b71f2c5c1a"} Oct 07 12:29:10 crc kubenswrapper[4702]: I1007 12:29:10.361363 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-86dwm" podStartSLOduration=1.828485414 podStartE2EDuration="4.361333262s" podCreationTimestamp="2025-10-07 12:29:06 +0000 UTC" firstStartedPulling="2025-10-07 12:29:07.295958238 +0000 UTC m=+303.471694354" lastFinishedPulling="2025-10-07 12:29:09.828806076 +0000 UTC m=+306.004542202" observedRunningTime="2025-10-07 12:29:10.360748174 +0000 UTC m=+306.536484290" watchObservedRunningTime="2025-10-07 12:29:10.361333262 +0000 UTC m=+306.537069378" Oct 07 12:29:11 crc kubenswrapper[4702]: I1007 12:29:11.344839 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2ntvb" event={"ID":"de46fdcb-23b3-4c75-baf4-584a8b29d3d3","Type":"ContainerStarted","Data":"e2c28633988c7ade192e2673b019c1bca8d0b86d5c07d95406a81208d671e13e"} Oct 07 12:29:12 crc kubenswrapper[4702]: I1007 12:29:12.378410 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2ntvb" podStartSLOduration=3.657700453 podStartE2EDuration="6.378386808s" podCreationTimestamp="2025-10-07 12:29:06 +0000 UTC" firstStartedPulling="2025-10-07 12:29:08.320000153 +0000 UTC m=+304.495736269" lastFinishedPulling="2025-10-07 12:29:11.040686508 +0000 UTC m=+307.216422624" observedRunningTime="2025-10-07 12:29:12.373915553 +0000 UTC m=+308.549651679" watchObservedRunningTime="2025-10-07 12:29:12.378386808 +0000 UTC m=+308.554122934" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.147923 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.148284 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.197170 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.355180 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.355336 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.399235 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.399435 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m4xtt" Oct 07 12:29:14 crc kubenswrapper[4702]: I1007 12:29:14.449879 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:29:16 crc kubenswrapper[4702]: I1007 12:29:16.685510 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:16 crc kubenswrapper[4702]: I1007 12:29:16.686056 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:16 crc kubenswrapper[4702]: I1007 12:29:16.744821 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:16 crc kubenswrapper[4702]: I1007 12:29:16.783440 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:16 crc kubenswrapper[4702]: I1007 12:29:16.783501 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:16 crc kubenswrapper[4702]: I1007 12:29:16.832008 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:17 crc kubenswrapper[4702]: I1007 12:29:17.416567 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2ntvb" Oct 07 12:29:17 crc kubenswrapper[4702]: I1007 12:29:17.418018 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-86dwm" Oct 07 12:29:39 crc kubenswrapper[4702]: I1007 12:29:39.263568 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:29:39 crc kubenswrapper[4702]: I1007 12:29:39.264599 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.167547 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh"] Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.169439 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.172710 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.172753 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.173424 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh"] Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.317675 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqkzc\" (UniqueName: \"kubernetes.io/projected/c9ef2928-8aab-4af4-9abf-9b75251898e0-kube-api-access-xqkzc\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.317743 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9ef2928-8aab-4af4-9abf-9b75251898e0-config-volume\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.317771 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9ef2928-8aab-4af4-9abf-9b75251898e0-secret-volume\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.418850 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqkzc\" (UniqueName: \"kubernetes.io/projected/c9ef2928-8aab-4af4-9abf-9b75251898e0-kube-api-access-xqkzc\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.418937 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9ef2928-8aab-4af4-9abf-9b75251898e0-config-volume\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.418962 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9ef2928-8aab-4af4-9abf-9b75251898e0-secret-volume\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.420208 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9ef2928-8aab-4af4-9abf-9b75251898e0-config-volume\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.428656 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9ef2928-8aab-4af4-9abf-9b75251898e0-secret-volume\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.441169 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqkzc\" (UniqueName: \"kubernetes.io/projected/c9ef2928-8aab-4af4-9abf-9b75251898e0-kube-api-access-xqkzc\") pod \"collect-profiles-29330670-t6vfh\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.496182 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:00 crc kubenswrapper[4702]: I1007 12:30:00.970157 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh"] Oct 07 12:30:00 crc kubenswrapper[4702]: W1007 12:30:00.983668 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ef2928_8aab_4af4_9abf_9b75251898e0.slice/crio-55c6bea73e845a133c8cb0cab397b0b87290e190a0b873648d44d365557e5e02 WatchSource:0}: Error finding container 55c6bea73e845a133c8cb0cab397b0b87290e190a0b873648d44d365557e5e02: Status 404 returned error can't find the container with id 55c6bea73e845a133c8cb0cab397b0b87290e190a0b873648d44d365557e5e02 Oct 07 12:30:01 crc kubenswrapper[4702]: I1007 12:30:01.676987 4702 generic.go:334] "Generic (PLEG): container finished" podID="c9ef2928-8aab-4af4-9abf-9b75251898e0" containerID="068ea61adae15fd2bb2fe4b59fca2a6a85c79171160c5553527c63d15a55b9c8" exitCode=0 Oct 07 12:30:01 crc kubenswrapper[4702]: I1007 12:30:01.677056 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" event={"ID":"c9ef2928-8aab-4af4-9abf-9b75251898e0","Type":"ContainerDied","Data":"068ea61adae15fd2bb2fe4b59fca2a6a85c79171160c5553527c63d15a55b9c8"} Oct 07 12:30:01 crc kubenswrapper[4702]: I1007 12:30:01.677099 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" event={"ID":"c9ef2928-8aab-4af4-9abf-9b75251898e0","Type":"ContainerStarted","Data":"55c6bea73e845a133c8cb0cab397b0b87290e190a0b873648d44d365557e5e02"} Oct 07 12:30:02 crc kubenswrapper[4702]: I1007 12:30:02.919160 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.063731 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9ef2928-8aab-4af4-9abf-9b75251898e0-secret-volume\") pod \"c9ef2928-8aab-4af4-9abf-9b75251898e0\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.063997 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9ef2928-8aab-4af4-9abf-9b75251898e0-config-volume\") pod \"c9ef2928-8aab-4af4-9abf-9b75251898e0\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.064125 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqkzc\" (UniqueName: \"kubernetes.io/projected/c9ef2928-8aab-4af4-9abf-9b75251898e0-kube-api-access-xqkzc\") pod \"c9ef2928-8aab-4af4-9abf-9b75251898e0\" (UID: \"c9ef2928-8aab-4af4-9abf-9b75251898e0\") " Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.065231 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9ef2928-8aab-4af4-9abf-9b75251898e0-config-volume" (OuterVolumeSpecName: "config-volume") pod "c9ef2928-8aab-4af4-9abf-9b75251898e0" (UID: "c9ef2928-8aab-4af4-9abf-9b75251898e0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.071631 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ef2928-8aab-4af4-9abf-9b75251898e0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c9ef2928-8aab-4af4-9abf-9b75251898e0" (UID: "c9ef2928-8aab-4af4-9abf-9b75251898e0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.072057 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ef2928-8aab-4af4-9abf-9b75251898e0-kube-api-access-xqkzc" (OuterVolumeSpecName: "kube-api-access-xqkzc") pod "c9ef2928-8aab-4af4-9abf-9b75251898e0" (UID: "c9ef2928-8aab-4af4-9abf-9b75251898e0"). InnerVolumeSpecName "kube-api-access-xqkzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.166179 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9ef2928-8aab-4af4-9abf-9b75251898e0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.166248 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqkzc\" (UniqueName: \"kubernetes.io/projected/c9ef2928-8aab-4af4-9abf-9b75251898e0-kube-api-access-xqkzc\") on node \"crc\" DevicePath \"\"" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.166275 4702 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9ef2928-8aab-4af4-9abf-9b75251898e0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.692324 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" event={"ID":"c9ef2928-8aab-4af4-9abf-9b75251898e0","Type":"ContainerDied","Data":"55c6bea73e845a133c8cb0cab397b0b87290e190a0b873648d44d365557e5e02"} Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.692398 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55c6bea73e845a133c8cb0cab397b0b87290e190a0b873648d44d365557e5e02" Oct 07 12:30:03 crc kubenswrapper[4702]: I1007 12:30:03.692497 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330670-t6vfh" Oct 07 12:30:09 crc kubenswrapper[4702]: I1007 12:30:09.262706 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:30:09 crc kubenswrapper[4702]: I1007 12:30:09.263668 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.263716 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.264794 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.264863 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.265709 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"446febc400f1513a5b71f983f1776c7a64b2c92e3c79409f2d677ce29b1b6901"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.265777 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://446febc400f1513a5b71f983f1776c7a64b2c92e3c79409f2d677ce29b1b6901" gracePeriod=600 Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.937011 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="446febc400f1513a5b71f983f1776c7a64b2c92e3c79409f2d677ce29b1b6901" exitCode=0 Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.937071 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"446febc400f1513a5b71f983f1776c7a64b2c92e3c79409f2d677ce29b1b6901"} Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.937887 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"dbea129310a361471f0e292025f60b439ffa2aae1c5e93aa00ef5e1ce0d3006d"} Oct 07 12:30:39 crc kubenswrapper[4702]: I1007 12:30:39.937928 4702 scope.go:117] "RemoveContainer" containerID="7683347af1e908e21c421fb6e793f2c1df1d6147766db7beb3c4d6aca7f23342" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.638475 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8qpds"] Oct 07 12:31:37 crc kubenswrapper[4702]: E1007 12:31:37.639543 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ef2928-8aab-4af4-9abf-9b75251898e0" containerName="collect-profiles" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.639559 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ef2928-8aab-4af4-9abf-9b75251898e0" containerName="collect-profiles" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.639655 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ef2928-8aab-4af4-9abf-9b75251898e0" containerName="collect-profiles" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.640418 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.659443 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8qpds"] Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734207 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61af08f9-dbf0-4886-abe5-0900a62272c1-registry-certificates\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734417 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61af08f9-dbf0-4886-abe5-0900a62272c1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734533 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-bound-sa-token\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734576 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61af08f9-dbf0-4886-abe5-0900a62272c1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734640 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61af08f9-dbf0-4886-abe5-0900a62272c1-trusted-ca\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734681 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-registry-tls\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734723 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.734746 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scm8r\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-kube-api-access-scm8r\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.799450 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836305 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61af08f9-dbf0-4886-abe5-0900a62272c1-trusted-ca\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836362 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-registry-tls\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836393 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scm8r\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-kube-api-access-scm8r\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836436 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61af08f9-dbf0-4886-abe5-0900a62272c1-registry-certificates\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836458 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61af08f9-dbf0-4886-abe5-0900a62272c1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836496 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-bound-sa-token\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.836513 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61af08f9-dbf0-4886-abe5-0900a62272c1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.837465 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61af08f9-dbf0-4886-abe5-0900a62272c1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.838185 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61af08f9-dbf0-4886-abe5-0900a62272c1-trusted-ca\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.838486 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61af08f9-dbf0-4886-abe5-0900a62272c1-registry-certificates\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.845876 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-registry-tls\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.846762 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61af08f9-dbf0-4886-abe5-0900a62272c1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.853967 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-bound-sa-token\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.855623 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scm8r\" (UniqueName: \"kubernetes.io/projected/61af08f9-dbf0-4886-abe5-0900a62272c1-kube-api-access-scm8r\") pod \"image-registry-66df7c8f76-8qpds\" (UID: \"61af08f9-dbf0-4886-abe5-0900a62272c1\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:37 crc kubenswrapper[4702]: I1007 12:31:37.960430 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:38 crc kubenswrapper[4702]: I1007 12:31:38.376011 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8qpds"] Oct 07 12:31:39 crc kubenswrapper[4702]: I1007 12:31:39.318391 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" event={"ID":"61af08f9-dbf0-4886-abe5-0900a62272c1","Type":"ContainerStarted","Data":"ef6a7a56e1af98cdcee8515ddedace4657a60bf96d56c6d827cdabe18f3df6f2"} Oct 07 12:31:39 crc kubenswrapper[4702]: I1007 12:31:39.319265 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:39 crc kubenswrapper[4702]: I1007 12:31:39.319351 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" event={"ID":"61af08f9-dbf0-4886-abe5-0900a62272c1","Type":"ContainerStarted","Data":"ec3cce282ee8074e7cf7eaa73fc6bd2feebbd9f2b088e239ebf7c7242c3e69d0"} Oct 07 12:31:39 crc kubenswrapper[4702]: I1007 12:31:39.342562 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" podStartSLOduration=2.342502505 podStartE2EDuration="2.342502505s" podCreationTimestamp="2025-10-07 12:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:31:39.339491017 +0000 UTC m=+455.515227213" watchObservedRunningTime="2025-10-07 12:31:39.342502505 +0000 UTC m=+455.518238671" Oct 07 12:31:57 crc kubenswrapper[4702]: I1007 12:31:57.969848 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8qpds" Oct 07 12:31:58 crc kubenswrapper[4702]: I1007 12:31:58.035205 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vw7pc"] Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.082680 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" podUID="e0a452b7-fe35-404d-8864-ff8d6be236da" containerName="registry" containerID="cri-o://6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c" gracePeriod=30 Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.505885 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.609582 4702 generic.go:334] "Generic (PLEG): container finished" podID="e0a452b7-fe35-404d-8864-ff8d6be236da" containerID="6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c" exitCode=0 Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.609681 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" event={"ID":"e0a452b7-fe35-404d-8864-ff8d6be236da","Type":"ContainerDied","Data":"6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c"} Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.609754 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" event={"ID":"e0a452b7-fe35-404d-8864-ff8d6be236da","Type":"ContainerDied","Data":"1758f1ae83b2ed479415c950c78426cf92a51402f394e06252acf0500baacf9e"} Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.609817 4702 scope.go:117] "RemoveContainer" containerID="6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.609828 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vw7pc" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.635149 4702 scope.go:117] "RemoveContainer" containerID="6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c" Oct 07 12:32:23 crc kubenswrapper[4702]: E1007 12:32:23.635720 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c\": container with ID starting with 6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c not found: ID does not exist" containerID="6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.635783 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c"} err="failed to get container status \"6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c\": rpc error: code = NotFound desc = could not find container \"6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c\": container with ID starting with 6220fde05685b3051728a9ef0ab0365fb66a365ffe8c36f1692051f211f2f88c not found: ID does not exist" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.685809 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.685888 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-certificates\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.685971 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e0a452b7-fe35-404d-8864-ff8d6be236da-installation-pull-secrets\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.686026 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e0a452b7-fe35-404d-8864-ff8d6be236da-ca-trust-extracted\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.686057 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-bound-sa-token\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.686106 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-trusted-ca\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.686153 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz5bw\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-kube-api-access-qz5bw\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.686214 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-tls\") pod \"e0a452b7-fe35-404d-8864-ff8d6be236da\" (UID: \"e0a452b7-fe35-404d-8864-ff8d6be236da\") " Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.687424 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.690519 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.693586 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a452b7-fe35-404d-8864-ff8d6be236da-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.693835 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.699394 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-kube-api-access-qz5bw" (OuterVolumeSpecName: "kube-api-access-qz5bw") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "kube-api-access-qz5bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.703374 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.707729 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.722757 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0a452b7-fe35-404d-8864-ff8d6be236da-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e0a452b7-fe35-404d-8864-ff8d6be236da" (UID: "e0a452b7-fe35-404d-8864-ff8d6be236da"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788064 4702 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e0a452b7-fe35-404d-8864-ff8d6be236da-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788115 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788132 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788146 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz5bw\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-kube-api-access-qz5bw\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788164 4702 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788178 4702 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e0a452b7-fe35-404d-8864-ff8d6be236da-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.788192 4702 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e0a452b7-fe35-404d-8864-ff8d6be236da-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.963028 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vw7pc"] Oct 07 12:32:23 crc kubenswrapper[4702]: I1007 12:32:23.968883 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vw7pc"] Oct 07 12:32:24 crc kubenswrapper[4702]: I1007 12:32:24.647396 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0a452b7-fe35-404d-8864-ff8d6be236da" path="/var/lib/kubelet/pods/e0a452b7-fe35-404d-8864-ff8d6be236da/volumes" Oct 07 12:32:39 crc kubenswrapper[4702]: I1007 12:32:39.263263 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:32:39 crc kubenswrapper[4702]: I1007 12:32:39.264260 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:33:09 crc kubenswrapper[4702]: I1007 12:33:09.262465 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:33:09 crc kubenswrapper[4702]: I1007 12:33:09.263585 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:33:39 crc kubenswrapper[4702]: I1007 12:33:39.263752 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:33:39 crc kubenswrapper[4702]: I1007 12:33:39.264758 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:33:39 crc kubenswrapper[4702]: I1007 12:33:39.264840 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:33:39 crc kubenswrapper[4702]: I1007 12:33:39.266031 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbea129310a361471f0e292025f60b439ffa2aae1c5e93aa00ef5e1ce0d3006d"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:33:39 crc kubenswrapper[4702]: I1007 12:33:39.266136 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://dbea129310a361471f0e292025f60b439ffa2aae1c5e93aa00ef5e1ce0d3006d" gracePeriod=600 Oct 07 12:33:40 crc kubenswrapper[4702]: I1007 12:33:40.170732 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="dbea129310a361471f0e292025f60b439ffa2aae1c5e93aa00ef5e1ce0d3006d" exitCode=0 Oct 07 12:33:40 crc kubenswrapper[4702]: I1007 12:33:40.170775 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"dbea129310a361471f0e292025f60b439ffa2aae1c5e93aa00ef5e1ce0d3006d"} Oct 07 12:33:40 crc kubenswrapper[4702]: I1007 12:33:40.171263 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"ef24fdfca720c04e2e8d0834767736417773f1bf6dd8396fbe53fc5a646c234f"} Oct 07 12:33:40 crc kubenswrapper[4702]: I1007 12:33:40.171298 4702 scope.go:117] "RemoveContainer" containerID="446febc400f1513a5b71f983f1776c7a64b2c92e3c79409f2d677ce29b1b6901" Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.290452 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pcg5c"] Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292231 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="sbdb" containerID="cri-o://b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292269 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="nbdb" containerID="cri-o://1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292337 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292406 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-acl-logging" containerID="cri-o://ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292410 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="northd" containerID="cri-o://0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292460 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-node" containerID="cri-o://28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.292144 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-controller" containerID="cri-o://422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.344527 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" containerID="cri-o://3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" gracePeriod=30 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.537436 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/3.log" Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.540553 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovn-acl-logging/0.log" Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.541357 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" exitCode=0 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.541396 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" exitCode=0 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.541408 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" exitCode=143 Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.541445 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.541510 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} Oct 07 12:34:30 crc kubenswrapper[4702]: I1007 12:34:30.541521 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.675169 4702 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 is running failed: container process not found" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.676121 4702 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 is running failed: container process not found" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.676112 4702 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 is running failed: container process not found" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.676654 4702 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 is running failed: container process not found" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.676669 4702 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 is running failed: container process not found" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.676690 4702 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="nbdb" Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.677093 4702 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 is running failed: container process not found" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 07 12:34:30 crc kubenswrapper[4702]: E1007 12:34:30.677147 4702 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="sbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.151061 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/3.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.156153 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovn-acl-logging/0.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.157196 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovn-controller/0.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.158194 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.253370 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wmb4n"] Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.253987 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kubecfg-setup" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254074 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kubecfg-setup" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254102 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254152 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254171 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254185 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254685 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-acl-logging" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254713 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-acl-logging" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254731 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="northd" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254743 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="northd" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254764 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="nbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254777 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="nbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254801 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a452b7-fe35-404d-8864-ff8d6be236da" containerName="registry" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254813 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a452b7-fe35-404d-8864-ff8d6be236da" containerName="registry" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254832 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="sbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254844 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="sbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254861 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-node" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254938 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-node" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254960 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254973 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.254987 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.254999 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.255014 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255026 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.255048 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255092 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255298 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-node" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255319 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255335 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255351 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255369 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255386 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-acl-logging" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255403 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="nbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255418 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="sbdb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255434 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="northd" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255451 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a452b7-fe35-404d-8864-ff8d6be236da" containerName="registry" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255466 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovn-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255482 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.255721 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.255739 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.256312 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerName="ovnkube-controller" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.262770 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324321 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-etc-openvswitch\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324404 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-var-lib-openvswitch\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324435 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-ovn-kubernetes\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324465 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-bin\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324507 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d2c9\" (UniqueName: \"kubernetes.io/projected/68b62ec2-6971-4e45-bff9-5b65b657b6b6-kube-api-access-6d2c9\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324513 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324552 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-systemd\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324680 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-slash\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324579 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324753 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324599 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324726 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-kubelet\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324814 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-slash" (OuterVolumeSpecName: "host-slash") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324859 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324859 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-systemd-units\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324980 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-env-overrides\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325097 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-ovn\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325162 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-script-lib\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325202 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325232 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-openvswitch\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325315 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-config\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325359 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-log-socket\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325400 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-netns\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325441 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-netd\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325469 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-log-socket" (OuterVolumeSpecName: "log-socket") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325465 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325498 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325498 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325561 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovn-node-metrics-cert\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325561 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325518 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325619 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-node-log\") pod \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\" (UID: \"68b62ec2-6971-4e45-bff9-5b65b657b6b6\") " Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325664 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325790 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-node-log" (OuterVolumeSpecName: "node-log") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.325959 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.324689 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326218 4702 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326253 4702 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326274 4702 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326288 4702 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-slash\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326303 4702 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326318 4702 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326333 4702 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326348 4702 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326363 4702 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-log-socket\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326378 4702 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326406 4702 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326424 4702 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.326422 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.334410 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b62ec2-6971-4e45-bff9-5b65b657b6b6-kube-api-access-6d2c9" (OuterVolumeSpecName: "kube-api-access-6d2c9") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "kube-api-access-6d2c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.341318 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.350523 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "68b62ec2-6971-4e45-bff9-5b65b657b6b6" (UID: "68b62ec2-6971-4e45-bff9-5b65b657b6b6"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.427571 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-ovn\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.427920 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-cni-netd\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428084 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-systemd\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428226 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-log-socket\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428324 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-etc-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428429 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpgz6\" (UniqueName: \"kubernetes.io/projected/6c7b68df-3109-49e2-b7df-0973d0514a19-kube-api-access-wpgz6\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428532 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c7b68df-3109-49e2-b7df-0973d0514a19-ovn-node-metrics-cert\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428633 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-env-overrides\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428730 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-kubelet\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.428822 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-systemd-units\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429037 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-run-netns\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429219 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-cni-bin\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429343 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-ovnkube-config\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429381 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-slash\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429441 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429649 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-run-ovn-kubernetes\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429782 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-var-lib-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429873 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-node-log\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.429995 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-ovnkube-script-lib\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430112 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430253 4702 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430332 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d2c9\" (UniqueName: \"kubernetes.io/projected/68b62ec2-6971-4e45-bff9-5b65b657b6b6-kube-api-access-6d2c9\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430397 4702 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430464 4702 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430526 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430594 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430661 4702 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b62ec2-6971-4e45-bff9-5b65b657b6b6-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.430721 4702 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68b62ec2-6971-4e45-bff9-5b65b657b6b6-node-log\") on node \"crc\" DevicePath \"\"" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531516 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-etc-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531570 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpgz6\" (UniqueName: \"kubernetes.io/projected/6c7b68df-3109-49e2-b7df-0973d0514a19-kube-api-access-wpgz6\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531600 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c7b68df-3109-49e2-b7df-0973d0514a19-ovn-node-metrics-cert\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531622 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-env-overrides\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531649 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-kubelet\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531670 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-systemd-units\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531697 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-run-netns\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531726 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-cni-bin\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531717 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-etc-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531757 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-ovnkube-config\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531866 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-slash\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.531959 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532009 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-run-ovn-kubernetes\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532098 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-var-lib-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532106 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-slash\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532154 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-node-log\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532222 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-ovnkube-script-lib\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532297 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532578 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-cni-netd\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532632 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-ovn\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532674 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-systemd\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532700 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-log-socket\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532702 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-ovnkube-config\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532752 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532830 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-ovn\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532832 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-run-netns\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532858 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-node-log\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532872 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-log-socket\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532912 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-cni-netd\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532877 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532839 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-cni-bin\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532937 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-kubelet\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532915 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-host-run-ovn-kubernetes\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.532998 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-systemd-units\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.533009 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-run-systemd\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.533011 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-env-overrides\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.533246 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c7b68df-3109-49e2-b7df-0973d0514a19-var-lib-openvswitch\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.533325 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6c7b68df-3109-49e2-b7df-0973d0514a19-ovnkube-script-lib\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.537437 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c7b68df-3109-49e2-b7df-0973d0514a19-ovn-node-metrics-cert\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.551803 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovnkube-controller/3.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.555084 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovn-acl-logging/0.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.555801 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pcg5c_68b62ec2-6971-4e45-bff9-5b65b657b6b6/ovn-controller/0.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556312 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" exitCode=0 Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556398 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556478 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556512 4702 scope.go:117] "RemoveContainer" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556542 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556422 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" exitCode=0 Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556683 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" exitCode=0 Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556780 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" exitCode=0 Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556856 4702 generic.go:334] "Generic (PLEG): container finished" podID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" exitCode=143 Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.556718 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557108 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557193 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557272 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pcg5c" event={"ID":"68b62ec2-6971-4e45-bff9-5b65b657b6b6","Type":"ContainerDied","Data":"3aa3931da2c247feeb5f4a2682d04beaac9294b0792af0a544809f6b69162c8c"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557335 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557398 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557452 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557504 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557551 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557590 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpgz6\" (UniqueName: \"kubernetes.io/projected/6c7b68df-3109-49e2-b7df-0973d0514a19-kube-api-access-wpgz6\") pod \"ovnkube-node-wmb4n\" (UID: \"6c7b68df-3109-49e2-b7df-0973d0514a19\") " pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557600 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557652 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557672 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.557681 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.561414 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/2.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.562194 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/1.log" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.562289 4702 generic.go:334] "Generic (PLEG): container finished" podID="943865b8-ec83-4c7e-9e71-b02eaea71c1d" containerID="33efb0b3bfe61433e0cea825895177a26cf9b61dbcc4e7bc8f53636d1f58bbfd" exitCode=2 Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.562326 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerDied","Data":"33efb0b3bfe61433e0cea825895177a26cf9b61dbcc4e7bc8f53636d1f58bbfd"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.562411 4702 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b"} Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.563095 4702 scope.go:117] "RemoveContainer" containerID="33efb0b3bfe61433e0cea825895177a26cf9b61dbcc4e7bc8f53636d1f58bbfd" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.563442 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9s7h_openshift-multus(943865b8-ec83-4c7e-9e71-b02eaea71c1d)\"" pod="openshift-multus/multus-h9s7h" podUID="943865b8-ec83-4c7e-9e71-b02eaea71c1d" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.583015 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.620606 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pcg5c"] Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.627696 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.629257 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pcg5c"] Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.666948 4702 scope.go:117] "RemoveContainer" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.682569 4702 scope.go:117] "RemoveContainer" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.710199 4702 scope.go:117] "RemoveContainer" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.733766 4702 scope.go:117] "RemoveContainer" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.751804 4702 scope.go:117] "RemoveContainer" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.768070 4702 scope.go:117] "RemoveContainer" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.786147 4702 scope.go:117] "RemoveContainer" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.858856 4702 scope.go:117] "RemoveContainer" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.873294 4702 scope.go:117] "RemoveContainer" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.873744 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": container with ID starting with 3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7 not found: ID does not exist" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.873791 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} err="failed to get container status \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": rpc error: code = NotFound desc = could not find container \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": container with ID starting with 3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.873823 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.874274 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": container with ID starting with 30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c not found: ID does not exist" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.874356 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} err="failed to get container status \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": rpc error: code = NotFound desc = could not find container \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": container with ID starting with 30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.874423 4702 scope.go:117] "RemoveContainer" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.874917 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": container with ID starting with b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 not found: ID does not exist" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.874968 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} err="failed to get container status \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": rpc error: code = NotFound desc = could not find container \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": container with ID starting with b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.874990 4702 scope.go:117] "RemoveContainer" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.875301 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": container with ID starting with 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 not found: ID does not exist" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.875369 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} err="failed to get container status \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": rpc error: code = NotFound desc = could not find container \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": container with ID starting with 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.875407 4702 scope.go:117] "RemoveContainer" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.875707 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": container with ID starting with 0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234 not found: ID does not exist" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.875742 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} err="failed to get container status \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": rpc error: code = NotFound desc = could not find container \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": container with ID starting with 0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.875763 4702 scope.go:117] "RemoveContainer" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.876289 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": container with ID starting with 9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0 not found: ID does not exist" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.876324 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} err="failed to get container status \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": rpc error: code = NotFound desc = could not find container \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": container with ID starting with 9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.876347 4702 scope.go:117] "RemoveContainer" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.876759 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": container with ID starting with 28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661 not found: ID does not exist" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.876789 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} err="failed to get container status \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": rpc error: code = NotFound desc = could not find container \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": container with ID starting with 28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.876807 4702 scope.go:117] "RemoveContainer" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.877552 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": container with ID starting with ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb not found: ID does not exist" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.877571 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} err="failed to get container status \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": rpc error: code = NotFound desc = could not find container \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": container with ID starting with ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.877583 4702 scope.go:117] "RemoveContainer" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.877779 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": container with ID starting with 422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486 not found: ID does not exist" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.877808 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} err="failed to get container status \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": rpc error: code = NotFound desc = could not find container \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": container with ID starting with 422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.877821 4702 scope.go:117] "RemoveContainer" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" Oct 07 12:34:31 crc kubenswrapper[4702]: E1007 12:34:31.878098 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": container with ID starting with 323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c not found: ID does not exist" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.878153 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c"} err="failed to get container status \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": rpc error: code = NotFound desc = could not find container \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": container with ID starting with 323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.878182 4702 scope.go:117] "RemoveContainer" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.878466 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} err="failed to get container status \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": rpc error: code = NotFound desc = could not find container \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": container with ID starting with 3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.878498 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.878740 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} err="failed to get container status \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": rpc error: code = NotFound desc = could not find container \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": container with ID starting with 30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.878779 4702 scope.go:117] "RemoveContainer" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.879127 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} err="failed to get container status \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": rpc error: code = NotFound desc = could not find container \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": container with ID starting with b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.879153 4702 scope.go:117] "RemoveContainer" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.879374 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} err="failed to get container status \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": rpc error: code = NotFound desc = could not find container \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": container with ID starting with 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.879405 4702 scope.go:117] "RemoveContainer" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.881155 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} err="failed to get container status \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": rpc error: code = NotFound desc = could not find container \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": container with ID starting with 0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.881200 4702 scope.go:117] "RemoveContainer" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.881664 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} err="failed to get container status \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": rpc error: code = NotFound desc = could not find container \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": container with ID starting with 9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.881683 4702 scope.go:117] "RemoveContainer" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.881928 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} err="failed to get container status \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": rpc error: code = NotFound desc = could not find container \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": container with ID starting with 28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.881946 4702 scope.go:117] "RemoveContainer" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.882195 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} err="failed to get container status \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": rpc error: code = NotFound desc = could not find container \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": container with ID starting with ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.882209 4702 scope.go:117] "RemoveContainer" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.882424 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} err="failed to get container status \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": rpc error: code = NotFound desc = could not find container \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": container with ID starting with 422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.882441 4702 scope.go:117] "RemoveContainer" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.882691 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c"} err="failed to get container status \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": rpc error: code = NotFound desc = could not find container \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": container with ID starting with 323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.882743 4702 scope.go:117] "RemoveContainer" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.883003 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} err="failed to get container status \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": rpc error: code = NotFound desc = could not find container \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": container with ID starting with 3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.883023 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.883777 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} err="failed to get container status \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": rpc error: code = NotFound desc = could not find container \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": container with ID starting with 30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.883797 4702 scope.go:117] "RemoveContainer" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.884069 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} err="failed to get container status \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": rpc error: code = NotFound desc = could not find container \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": container with ID starting with b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.884093 4702 scope.go:117] "RemoveContainer" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.884331 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} err="failed to get container status \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": rpc error: code = NotFound desc = could not find container \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": container with ID starting with 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.884350 4702 scope.go:117] "RemoveContainer" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.884719 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} err="failed to get container status \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": rpc error: code = NotFound desc = could not find container \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": container with ID starting with 0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.884738 4702 scope.go:117] "RemoveContainer" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885083 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} err="failed to get container status \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": rpc error: code = NotFound desc = could not find container \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": container with ID starting with 9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885105 4702 scope.go:117] "RemoveContainer" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885325 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} err="failed to get container status \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": rpc error: code = NotFound desc = could not find container \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": container with ID starting with 28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885343 4702 scope.go:117] "RemoveContainer" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885718 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} err="failed to get container status \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": rpc error: code = NotFound desc = could not find container \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": container with ID starting with ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885731 4702 scope.go:117] "RemoveContainer" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885940 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} err="failed to get container status \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": rpc error: code = NotFound desc = could not find container \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": container with ID starting with 422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.885959 4702 scope.go:117] "RemoveContainer" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.886412 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c"} err="failed to get container status \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": rpc error: code = NotFound desc = could not find container \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": container with ID starting with 323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.886434 4702 scope.go:117] "RemoveContainer" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.886687 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} err="failed to get container status \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": rpc error: code = NotFound desc = could not find container \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": container with ID starting with 3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.886711 4702 scope.go:117] "RemoveContainer" containerID="30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.886885 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c"} err="failed to get container status \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": rpc error: code = NotFound desc = could not find container \"30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c\": container with ID starting with 30fce8235fc8d660b6cd1c7d17ca86af99e37409573ea4f9d73b2e7e4e6cc97c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.886924 4702 scope.go:117] "RemoveContainer" containerID="b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887151 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98"} err="failed to get container status \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": rpc error: code = NotFound desc = could not find container \"b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98\": container with ID starting with b3db3b4d3f83076d43995a11002f15edecfadf42bb8d4e69eb9b4c9b9d46ae98 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887177 4702 scope.go:117] "RemoveContainer" containerID="1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887386 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2"} err="failed to get container status \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": rpc error: code = NotFound desc = could not find container \"1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2\": container with ID starting with 1dacf5ce44ab685a7dc244bc333d63c94c3db4d996b74cdb2b4f9757b0c606e2 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887409 4702 scope.go:117] "RemoveContainer" containerID="0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887622 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234"} err="failed to get container status \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": rpc error: code = NotFound desc = could not find container \"0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234\": container with ID starting with 0cf3b40cec1e067adea1c013272dea380256fec556e893abdef5136fb9ff2234 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887645 4702 scope.go:117] "RemoveContainer" containerID="9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887957 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0"} err="failed to get container status \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": rpc error: code = NotFound desc = could not find container \"9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0\": container with ID starting with 9fe6a68b34e6dcb119cd0952878997676deb001151d20a3fb97da7385d7c80b0 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.887981 4702 scope.go:117] "RemoveContainer" containerID="28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888240 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661"} err="failed to get container status \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": rpc error: code = NotFound desc = could not find container \"28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661\": container with ID starting with 28554d9b86b158da07932732afdece4d76a8baeabfe98f58dbdcb78224383661 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888258 4702 scope.go:117] "RemoveContainer" containerID="ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888480 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb"} err="failed to get container status \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": rpc error: code = NotFound desc = could not find container \"ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb\": container with ID starting with ccf2f77e08f6b5b034efe93e992d88ccbbaa91f536b104f5f4be8cd60aa08adb not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888505 4702 scope.go:117] "RemoveContainer" containerID="422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888722 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486"} err="failed to get container status \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": rpc error: code = NotFound desc = could not find container \"422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486\": container with ID starting with 422f31c3f4a60459db2ad42829efc288ae48c4f258d2e95a062092457e617486 not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888747 4702 scope.go:117] "RemoveContainer" containerID="323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.888996 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c"} err="failed to get container status \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": rpc error: code = NotFound desc = could not find container \"323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c\": container with ID starting with 323a2d2913e2c4f40cb736dba2663476073ed9a5269c548023270c3774024c5c not found: ID does not exist" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.889020 4702 scope.go:117] "RemoveContainer" containerID="3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7" Oct 07 12:34:31 crc kubenswrapper[4702]: I1007 12:34:31.889247 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7"} err="failed to get container status \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": rpc error: code = NotFound desc = could not find container \"3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7\": container with ID starting with 3d48c04b150f440140c96d4ffffa48b2bbe3ea2830839a523b05b3fd02eddfc7 not found: ID does not exist" Oct 07 12:34:32 crc kubenswrapper[4702]: I1007 12:34:32.571518 4702 generic.go:334] "Generic (PLEG): container finished" podID="6c7b68df-3109-49e2-b7df-0973d0514a19" containerID="210aeb82a137597ec3300e942c277fafde1920fc4c2f4ce5afd537a9e5f889cd" exitCode=0 Oct 07 12:34:32 crc kubenswrapper[4702]: I1007 12:34:32.571648 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerDied","Data":"210aeb82a137597ec3300e942c277fafde1920fc4c2f4ce5afd537a9e5f889cd"} Oct 07 12:34:32 crc kubenswrapper[4702]: I1007 12:34:32.573860 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"2d12fdedb3b0f619f5851e6f5b1f9ddb1ec3b06f9b87c75d7d211a5ce297a648"} Oct 07 12:34:32 crc kubenswrapper[4702]: I1007 12:34:32.649757 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b62ec2-6971-4e45-bff9-5b65b657b6b6" path="/var/lib/kubelet/pods/68b62ec2-6971-4e45-bff9-5b65b657b6b6/volumes" Oct 07 12:34:33 crc kubenswrapper[4702]: I1007 12:34:33.590020 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"b3438cf011810ac4f1381a8190f1d3181fd0447a26b4a7b4fa15327184bdc045"} Oct 07 12:34:33 crc kubenswrapper[4702]: I1007 12:34:33.590619 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"aebae83524c71f32839a1e1b7bdc52980d8b64d38ad710796d66c4987947ee70"} Oct 07 12:34:33 crc kubenswrapper[4702]: I1007 12:34:33.590653 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"52f8fe5b913e3f24ee9076b27e075403e8520eff4673f58c650b2b94223f38f1"} Oct 07 12:34:33 crc kubenswrapper[4702]: I1007 12:34:33.590674 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"637f99ab47d78fde415e7d1941cb1b0c06cbfdc111520da69f38ff182b7bc0ab"} Oct 07 12:34:33 crc kubenswrapper[4702]: I1007 12:34:33.590723 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"ca0b5928657207caf79c326a5b2889080c5fd74e6b7e8c1de2454684d7319681"} Oct 07 12:34:34 crc kubenswrapper[4702]: I1007 12:34:34.599595 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"c12b51066a27691a94d8a43d836acc38b58252cc37cbe44f6dbd87c03b64b1ba"} Oct 07 12:34:37 crc kubenswrapper[4702]: I1007 12:34:37.625985 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"2ce87e8b6017e4f340d0cb8bdb063c0238b25d33d450f7e35fbbd0dc771baca9"} Oct 07 12:34:38 crc kubenswrapper[4702]: I1007 12:34:38.638932 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" event={"ID":"6c7b68df-3109-49e2-b7df-0973d0514a19","Type":"ContainerStarted","Data":"f14106e3fdc9313ed1dd9b63d6a346e45dcab889bafd300841d3170375d9647f"} Oct 07 12:34:38 crc kubenswrapper[4702]: I1007 12:34:38.639448 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:38 crc kubenswrapper[4702]: I1007 12:34:38.639488 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:38 crc kubenswrapper[4702]: I1007 12:34:38.684048 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:38 crc kubenswrapper[4702]: I1007 12:34:38.689060 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" podStartSLOduration=7.689037059 podStartE2EDuration="7.689037059s" podCreationTimestamp="2025-10-07 12:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:34:38.687963068 +0000 UTC m=+634.863699234" watchObservedRunningTime="2025-10-07 12:34:38.689037059 +0000 UTC m=+634.864773175" Oct 07 12:34:39 crc kubenswrapper[4702]: I1007 12:34:39.647573 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:39 crc kubenswrapper[4702]: I1007 12:34:39.695490 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:34:46 crc kubenswrapper[4702]: I1007 12:34:46.641857 4702 scope.go:117] "RemoveContainer" containerID="33efb0b3bfe61433e0cea825895177a26cf9b61dbcc4e7bc8f53636d1f58bbfd" Oct 07 12:34:46 crc kubenswrapper[4702]: E1007 12:34:46.643521 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9s7h_openshift-multus(943865b8-ec83-4c7e-9e71-b02eaea71c1d)\"" pod="openshift-multus/multus-h9s7h" podUID="943865b8-ec83-4c7e-9e71-b02eaea71c1d" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.564821 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j"] Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.568190 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.571671 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.584561 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j"] Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.746142 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.746233 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.746264 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7bpz\" (UniqueName: \"kubernetes.io/projected/d93c1009-09dc-4db3-b49a-6981a02756be-kube-api-access-g7bpz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.848125 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.848199 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.848230 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7bpz\" (UniqueName: \"kubernetes.io/projected/d93c1009-09dc-4db3-b49a-6981a02756be-kube-api-access-g7bpz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.848836 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.849191 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.881747 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7bpz\" (UniqueName: \"kubernetes.io/projected/d93c1009-09dc-4db3-b49a-6981a02756be-kube-api-access-g7bpz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: I1007 12:34:56.904325 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: E1007 12:34:56.939429 4702 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(4f7efd404047b6814d56c9880241ed7ca41fe6c8c9b44a4adef812eb0b780377): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 12:34:56 crc kubenswrapper[4702]: E1007 12:34:56.939542 4702 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(4f7efd404047b6814d56c9880241ed7ca41fe6c8c9b44a4adef812eb0b780377): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: E1007 12:34:56.939577 4702 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(4f7efd404047b6814d56c9880241ed7ca41fe6c8c9b44a4adef812eb0b780377): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:56 crc kubenswrapper[4702]: E1007 12:34:56.939658 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace(d93c1009-09dc-4db3-b49a-6981a02756be)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace(d93c1009-09dc-4db3-b49a-6981a02756be)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(4f7efd404047b6814d56c9880241ed7ca41fe6c8c9b44a4adef812eb0b780377): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" Oct 07 12:34:57 crc kubenswrapper[4702]: I1007 12:34:57.843377 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:57 crc kubenswrapper[4702]: I1007 12:34:57.844380 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:57 crc kubenswrapper[4702]: E1007 12:34:57.896841 4702 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(7fca981d8ab46392af55dc0b2ca8d4a50fbc0bb15844d0d3654e9ebdbbca59e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 12:34:57 crc kubenswrapper[4702]: E1007 12:34:57.896991 4702 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(7fca981d8ab46392af55dc0b2ca8d4a50fbc0bb15844d0d3654e9ebdbbca59e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:57 crc kubenswrapper[4702]: E1007 12:34:57.897033 4702 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(7fca981d8ab46392af55dc0b2ca8d4a50fbc0bb15844d0d3654e9ebdbbca59e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:34:57 crc kubenswrapper[4702]: E1007 12:34:57.897119 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace(d93c1009-09dc-4db3-b49a-6981a02756be)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace(d93c1009-09dc-4db3-b49a-6981a02756be)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_openshift-marketplace_d93c1009-09dc-4db3-b49a-6981a02756be_0(7fca981d8ab46392af55dc0b2ca8d4a50fbc0bb15844d0d3654e9ebdbbca59e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" Oct 07 12:34:58 crc kubenswrapper[4702]: I1007 12:34:58.642100 4702 scope.go:117] "RemoveContainer" containerID="33efb0b3bfe61433e0cea825895177a26cf9b61dbcc4e7bc8f53636d1f58bbfd" Oct 07 12:34:58 crc kubenswrapper[4702]: I1007 12:34:58.857523 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/2.log" Oct 07 12:34:58 crc kubenswrapper[4702]: I1007 12:34:58.858888 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/1.log" Oct 07 12:34:59 crc kubenswrapper[4702]: I1007 12:34:59.871685 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/2.log" Oct 07 12:34:59 crc kubenswrapper[4702]: I1007 12:34:59.872624 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/1.log" Oct 07 12:34:59 crc kubenswrapper[4702]: I1007 12:34:59.872718 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9s7h" event={"ID":"943865b8-ec83-4c7e-9e71-b02eaea71c1d","Type":"ContainerStarted","Data":"ac7829f7786ab860e97e6bf576da72ec72cbb5e04b2c88aaa7860b4ba47c091b"} Oct 07 12:35:01 crc kubenswrapper[4702]: I1007 12:35:01.625505 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wmb4n" Oct 07 12:35:04 crc kubenswrapper[4702]: I1007 12:35:04.875086 4702 scope.go:117] "RemoveContainer" containerID="c7ca8c3e69a1289c79b531a84d511fb3812a5f1eaf485e80abe282df2985af8b" Oct 07 12:35:05 crc kubenswrapper[4702]: I1007 12:35:05.915714 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9s7h_943865b8-ec83-4c7e-9e71-b02eaea71c1d/kube-multus/2.log" Oct 07 12:35:11 crc kubenswrapper[4702]: I1007 12:35:11.640394 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:35:11 crc kubenswrapper[4702]: I1007 12:35:11.642306 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:35:12 crc kubenswrapper[4702]: I1007 12:35:12.142492 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j"] Oct 07 12:35:12 crc kubenswrapper[4702]: I1007 12:35:12.963284 4702 generic.go:334] "Generic (PLEG): container finished" podID="d93c1009-09dc-4db3-b49a-6981a02756be" containerID="0f43c6f289eeeb17c57c546051a8f325df7c6f915b0eece108a63d08c2f8a52d" exitCode=0 Oct 07 12:35:12 crc kubenswrapper[4702]: I1007 12:35:12.963337 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" event={"ID":"d93c1009-09dc-4db3-b49a-6981a02756be","Type":"ContainerDied","Data":"0f43c6f289eeeb17c57c546051a8f325df7c6f915b0eece108a63d08c2f8a52d"} Oct 07 12:35:12 crc kubenswrapper[4702]: I1007 12:35:12.963368 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" event={"ID":"d93c1009-09dc-4db3-b49a-6981a02756be","Type":"ContainerStarted","Data":"7dfdd55dbe685a17b585818a91402c666733745c91a1e7e30bd2de61a6d77210"} Oct 07 12:35:12 crc kubenswrapper[4702]: I1007 12:35:12.965917 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 12:35:15 crc kubenswrapper[4702]: I1007 12:35:15.985960 4702 generic.go:334] "Generic (PLEG): container finished" podID="d93c1009-09dc-4db3-b49a-6981a02756be" containerID="0317c2365e49e2e7349b9b30bfc666c7f5b6f78c6b44d7f265ed928cf9202b05" exitCode=0 Oct 07 12:35:15 crc kubenswrapper[4702]: I1007 12:35:15.986096 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" event={"ID":"d93c1009-09dc-4db3-b49a-6981a02756be","Type":"ContainerDied","Data":"0317c2365e49e2e7349b9b30bfc666c7f5b6f78c6b44d7f265ed928cf9202b05"} Oct 07 12:35:16 crc kubenswrapper[4702]: I1007 12:35:16.996374 4702 generic.go:334] "Generic (PLEG): container finished" podID="d93c1009-09dc-4db3-b49a-6981a02756be" containerID="83b4f0bad2c68371295c10b962fd326039519ea97388970b76048e60c1e600e3" exitCode=0 Oct 07 12:35:16 crc kubenswrapper[4702]: I1007 12:35:16.996442 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" event={"ID":"d93c1009-09dc-4db3-b49a-6981a02756be","Type":"ContainerDied","Data":"83b4f0bad2c68371295c10b962fd326039519ea97388970b76048e60c1e600e3"} Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.263100 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.462782 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7bpz\" (UniqueName: \"kubernetes.io/projected/d93c1009-09dc-4db3-b49a-6981a02756be-kube-api-access-g7bpz\") pod \"d93c1009-09dc-4db3-b49a-6981a02756be\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.462825 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-util\") pod \"d93c1009-09dc-4db3-b49a-6981a02756be\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.462876 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-bundle\") pod \"d93c1009-09dc-4db3-b49a-6981a02756be\" (UID: \"d93c1009-09dc-4db3-b49a-6981a02756be\") " Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.464998 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-bundle" (OuterVolumeSpecName: "bundle") pod "d93c1009-09dc-4db3-b49a-6981a02756be" (UID: "d93c1009-09dc-4db3-b49a-6981a02756be"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.472539 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93c1009-09dc-4db3-b49a-6981a02756be-kube-api-access-g7bpz" (OuterVolumeSpecName: "kube-api-access-g7bpz") pod "d93c1009-09dc-4db3-b49a-6981a02756be" (UID: "d93c1009-09dc-4db3-b49a-6981a02756be"). InnerVolumeSpecName "kube-api-access-g7bpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.485645 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-util" (OuterVolumeSpecName: "util") pod "d93c1009-09dc-4db3-b49a-6981a02756be" (UID: "d93c1009-09dc-4db3-b49a-6981a02756be"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.564602 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7bpz\" (UniqueName: \"kubernetes.io/projected/d93c1009-09dc-4db3-b49a-6981a02756be-kube-api-access-g7bpz\") on node \"crc\" DevicePath \"\"" Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.564661 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:35:18 crc kubenswrapper[4702]: I1007 12:35:18.564680 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c1009-09dc-4db3-b49a-6981a02756be-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:35:19 crc kubenswrapper[4702]: I1007 12:35:19.012937 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" event={"ID":"d93c1009-09dc-4db3-b49a-6981a02756be","Type":"ContainerDied","Data":"7dfdd55dbe685a17b585818a91402c666733745c91a1e7e30bd2de61a6d77210"} Oct 07 12:35:19 crc kubenswrapper[4702]: I1007 12:35:19.013005 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dfdd55dbe685a17b585818a91402c666733745c91a1e7e30bd2de61a6d77210" Oct 07 12:35:19 crc kubenswrapper[4702]: I1007 12:35:19.013049 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.884473 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf"] Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.885691 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="extract" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.885716 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="extract" Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.885731 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="pull" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.885741 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="pull" Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.885757 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="util" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.885766 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="util" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.885913 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93c1009-09dc-4db3-b49a-6981a02756be" containerName="extract" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.886467 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:29 crc kubenswrapper[4702]: W1007 12:35:29.893099 4702 reflector.go:561] object-"metallb-system"/"manager-account-dockercfg-vdfrq": failed to list *v1.Secret: secrets "manager-account-dockercfg-vdfrq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.893161 4702 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"manager-account-dockercfg-vdfrq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"manager-account-dockercfg-vdfrq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 12:35:29 crc kubenswrapper[4702]: W1007 12:35:29.893250 4702 reflector.go:561] object-"metallb-system"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.893262 4702 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 12:35:29 crc kubenswrapper[4702]: W1007 12:35:29.893283 4702 reflector.go:561] object-"metallb-system"/"metallb-operator-controller-manager-service-cert": failed to list *v1.Secret: secrets "metallb-operator-controller-manager-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 07 12:35:29 crc kubenswrapper[4702]: W1007 12:35:29.893335 4702 reflector.go:561] object-"metallb-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.893347 4702 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.893349 4702 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-controller-manager-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-controller-manager-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 12:35:29 crc kubenswrapper[4702]: W1007 12:35:29.893303 4702 reflector.go:561] object-"metallb-system"/"metallb-operator-webhook-server-cert": failed to list *v1.Secret: secrets "metallb-operator-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 07 12:35:29 crc kubenswrapper[4702]: E1007 12:35:29.893382 4702 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 12:35:29 crc kubenswrapper[4702]: I1007 12:35:29.917301 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf"] Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.022317 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h8vg\" (UniqueName: \"kubernetes.io/projected/10a60231-0c5a-463a-8836-310df99ffe9d-kube-api-access-9h8vg\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.022400 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-webhook-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.022426 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-apiservice-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.124048 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-webhook-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.124109 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-apiservice-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.124159 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h8vg\" (UniqueName: \"kubernetes.io/projected/10a60231-0c5a-463a-8836-310df99ffe9d-kube-api-access-9h8vg\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.147631 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8"] Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.148343 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.150584 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.151945 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.154090 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-9d4t7" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.171784 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8"] Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.327578 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/190a6066-4fa2-4dbe-bc52-6125fad41859-webhook-cert\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.327649 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-589pr\" (UniqueName: \"kubernetes.io/projected/190a6066-4fa2-4dbe-bc52-6125fad41859-kube-api-access-589pr\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.327685 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/190a6066-4fa2-4dbe-bc52-6125fad41859-apiservice-cert\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.428966 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/190a6066-4fa2-4dbe-bc52-6125fad41859-apiservice-cert\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.429120 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/190a6066-4fa2-4dbe-bc52-6125fad41859-webhook-cert\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.429158 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-589pr\" (UniqueName: \"kubernetes.io/projected/190a6066-4fa2-4dbe-bc52-6125fad41859-kube-api-access-589pr\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.437970 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/190a6066-4fa2-4dbe-bc52-6125fad41859-webhook-cert\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.444711 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/190a6066-4fa2-4dbe-bc52-6125fad41859-apiservice-cert\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:30 crc kubenswrapper[4702]: I1007 12:35:30.949630 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.038144 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.125303 4702 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.125326 4702 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.125453 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-apiservice-cert podName:10a60231-0c5a-463a-8836-310df99ffe9d nodeName:}" failed. No retries permitted until 2025-10-07 12:35:31.625422504 +0000 UTC m=+687.801158620 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-apiservice-cert") pod "metallb-operator-controller-manager-6fdccd459c-wjxtf" (UID: "10a60231-0c5a-463a-8836-310df99ffe9d") : failed to sync secret cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.125673 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-webhook-cert podName:10a60231-0c5a-463a-8836-310df99ffe9d nodeName:}" failed. No retries permitted until 2025-10-07 12:35:31.62562928 +0000 UTC m=+687.801365396 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-webhook-cert") pod "metallb-operator-controller-manager-6fdccd459c-wjxtf" (UID: "10a60231-0c5a-463a-8836-310df99ffe9d") : failed to sync secret cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.165477 4702 projected.go:288] Couldn't get configMap metallb-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.165553 4702 projected.go:194] Error preparing data for projected volume kube-api-access-9h8vg for pod metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf: failed to sync configmap cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.165635 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10a60231-0c5a-463a-8836-310df99ffe9d-kube-api-access-9h8vg podName:10a60231-0c5a-463a-8836-310df99ffe9d nodeName:}" failed. No retries permitted until 2025-10-07 12:35:31.665613798 +0000 UTC m=+687.841349914 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9h8vg" (UniqueName: "kubernetes.io/projected/10a60231-0c5a-463a-8836-310df99ffe9d-kube-api-access-9h8vg") pod "metallb-operator-controller-manager-6fdccd459c-wjxtf" (UID: "10a60231-0c5a-463a-8836-310df99ffe9d") : failed to sync configmap cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.229010 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-vdfrq" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.405198 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.441833 4702 projected.go:288] Couldn't get configMap metallb-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.441964 4702 projected.go:194] Error preparing data for projected volume kube-api-access-589pr for pod metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8: failed to sync configmap cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: E1007 12:35:31.442073 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/190a6066-4fa2-4dbe-bc52-6125fad41859-kube-api-access-589pr podName:190a6066-4fa2-4dbe-bc52-6125fad41859 nodeName:}" failed. No retries permitted until 2025-10-07 12:35:31.942042835 +0000 UTC m=+688.117778971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-589pr" (UniqueName: "kubernetes.io/projected/190a6066-4fa2-4dbe-bc52-6125fad41859-kube-api-access-589pr") pod "metallb-operator-webhook-server-5966c949c-9xjc8" (UID: "190a6066-4fa2-4dbe-bc52-6125fad41859") : failed to sync configmap cache: timed out waiting for the condition Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.489041 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.646093 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-webhook-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.646274 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-apiservice-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.654286 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-apiservice-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.657291 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10a60231-0c5a-463a-8836-310df99ffe9d-webhook-cert\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.747412 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h8vg\" (UniqueName: \"kubernetes.io/projected/10a60231-0c5a-463a-8836-310df99ffe9d-kube-api-access-9h8vg\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.754223 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h8vg\" (UniqueName: \"kubernetes.io/projected/10a60231-0c5a-463a-8836-310df99ffe9d-kube-api-access-9h8vg\") pod \"metallb-operator-controller-manager-6fdccd459c-wjxtf\" (UID: \"10a60231-0c5a-463a-8836-310df99ffe9d\") " pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.950874 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-589pr\" (UniqueName: \"kubernetes.io/projected/190a6066-4fa2-4dbe-bc52-6125fad41859-kube-api-access-589pr\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.959339 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-589pr\" (UniqueName: \"kubernetes.io/projected/190a6066-4fa2-4dbe-bc52-6125fad41859-kube-api-access-589pr\") pod \"metallb-operator-webhook-server-5966c949c-9xjc8\" (UID: \"190a6066-4fa2-4dbe-bc52-6125fad41859\") " pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:31 crc kubenswrapper[4702]: I1007 12:35:31.963453 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:32 crc kubenswrapper[4702]: I1007 12:35:32.006639 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:32 crc kubenswrapper[4702]: I1007 12:35:32.315804 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8"] Oct 07 12:35:32 crc kubenswrapper[4702]: W1007 12:35:32.323524 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod190a6066_4fa2_4dbe_bc52_6125fad41859.slice/crio-826c026850b09f694b1be236cc07c6299043f6d126f55ebce5eb261d594704e2 WatchSource:0}: Error finding container 826c026850b09f694b1be236cc07c6299043f6d126f55ebce5eb261d594704e2: Status 404 returned error can't find the container with id 826c026850b09f694b1be236cc07c6299043f6d126f55ebce5eb261d594704e2 Oct 07 12:35:32 crc kubenswrapper[4702]: I1007 12:35:32.374233 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf"] Oct 07 12:35:32 crc kubenswrapper[4702]: W1007 12:35:32.380159 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10a60231_0c5a_463a_8836_310df99ffe9d.slice/crio-ab7bcbace3cbcbdcab673303349f556891635cb3f350320825e08b15bf57ce3a WatchSource:0}: Error finding container ab7bcbace3cbcbdcab673303349f556891635cb3f350320825e08b15bf57ce3a: Status 404 returned error can't find the container with id ab7bcbace3cbcbdcab673303349f556891635cb3f350320825e08b15bf57ce3a Oct 07 12:35:33 crc kubenswrapper[4702]: I1007 12:35:33.101599 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" event={"ID":"10a60231-0c5a-463a-8836-310df99ffe9d","Type":"ContainerStarted","Data":"ab7bcbace3cbcbdcab673303349f556891635cb3f350320825e08b15bf57ce3a"} Oct 07 12:35:33 crc kubenswrapper[4702]: I1007 12:35:33.103612 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" event={"ID":"190a6066-4fa2-4dbe-bc52-6125fad41859","Type":"ContainerStarted","Data":"826c026850b09f694b1be236cc07c6299043f6d126f55ebce5eb261d594704e2"} Oct 07 12:35:38 crc kubenswrapper[4702]: I1007 12:35:38.138402 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" event={"ID":"10a60231-0c5a-463a-8836-310df99ffe9d","Type":"ContainerStarted","Data":"ecb60054beb69242eb367f932eb471b139d05b7d7bf4794725958caab39d1cf9"} Oct 07 12:35:38 crc kubenswrapper[4702]: I1007 12:35:38.139243 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:35:38 crc kubenswrapper[4702]: I1007 12:35:38.140923 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" event={"ID":"190a6066-4fa2-4dbe-bc52-6125fad41859","Type":"ContainerStarted","Data":"f5470a4dcce97c4e3ccf33af739cbf21dba8b5ee50445334f7cfa7498f6b664c"} Oct 07 12:35:38 crc kubenswrapper[4702]: I1007 12:35:38.141108 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:35:38 crc kubenswrapper[4702]: I1007 12:35:38.195509 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" podStartSLOduration=3.134114935 podStartE2EDuration="8.195480996s" podCreationTimestamp="2025-10-07 12:35:30 +0000 UTC" firstStartedPulling="2025-10-07 12:35:32.326591021 +0000 UTC m=+688.502327137" lastFinishedPulling="2025-10-07 12:35:37.387957082 +0000 UTC m=+693.563693198" observedRunningTime="2025-10-07 12:35:38.19320801 +0000 UTC m=+694.368944126" watchObservedRunningTime="2025-10-07 12:35:38.195480996 +0000 UTC m=+694.371217122" Oct 07 12:35:38 crc kubenswrapper[4702]: I1007 12:35:38.197564 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" podStartSLOduration=4.205313327 podStartE2EDuration="9.197555467s" podCreationTimestamp="2025-10-07 12:35:29 +0000 UTC" firstStartedPulling="2025-10-07 12:35:32.383131234 +0000 UTC m=+688.558867350" lastFinishedPulling="2025-10-07 12:35:37.375373374 +0000 UTC m=+693.551109490" observedRunningTime="2025-10-07 12:35:38.173832914 +0000 UTC m=+694.349569080" watchObservedRunningTime="2025-10-07 12:35:38.197555467 +0000 UTC m=+694.373291583" Oct 07 12:35:39 crc kubenswrapper[4702]: I1007 12:35:39.262407 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:35:39 crc kubenswrapper[4702]: I1007 12:35:39.263041 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:35:51 crc kubenswrapper[4702]: I1007 12:35:51.968732 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5966c949c-9xjc8" Oct 07 12:36:09 crc kubenswrapper[4702]: I1007 12:36:09.263031 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:36:09 crc kubenswrapper[4702]: I1007 12:36:09.264260 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.010216 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6fdccd459c-wjxtf" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.799859 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8w262"] Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.801066 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.803545 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6nskv"] Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.805800 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-76tg5" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.806203 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.807105 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.810864 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.821443 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8w262"] Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.824124 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.910104 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7vt7t"] Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.911092 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7vt7t" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.913582 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.913593 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.913810 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-gbvkh" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.913836 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.918634 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-reloader\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.918831 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f9fc314-be24-41d6-80bd-e4622cd1804e-metrics-certs\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919038 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9pvm\" (UniqueName: \"kubernetes.io/projected/3f9fc314-be24-41d6-80bd-e4622cd1804e-kube-api-access-g9pvm\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919149 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-metrics\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919332 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-conf\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919429 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-sockets\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919540 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26b44d11-5310-42a7-bbbc-fc95de8743a2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8w262\" (UID: \"26b44d11-5310-42a7-bbbc-fc95de8743a2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919655 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bkp8\" (UniqueName: \"kubernetes.io/projected/26b44d11-5310-42a7-bbbc-fc95de8743a2-kube-api-access-5bkp8\") pod \"frr-k8s-webhook-server-64bf5d555-8w262\" (UID: \"26b44d11-5310-42a7-bbbc-fc95de8743a2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.919753 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-startup\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.926166 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-bwfk4"] Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.928381 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.931071 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 07 12:36:12 crc kubenswrapper[4702]: I1007 12:36:12.947992 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-bwfk4"] Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.020957 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26b44d11-5310-42a7-bbbc-fc95de8743a2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8w262\" (UID: \"26b44d11-5310-42a7-bbbc-fc95de8743a2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021048 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bkp8\" (UniqueName: \"kubernetes.io/projected/26b44d11-5310-42a7-bbbc-fc95de8743a2-kube-api-access-5bkp8\") pod \"frr-k8s-webhook-server-64bf5d555-8w262\" (UID: \"26b44d11-5310-42a7-bbbc-fc95de8743a2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021075 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-startup\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021103 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-reloader\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021122 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f9fc314-be24-41d6-80bd-e4622cd1804e-metrics-certs\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021145 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs5mp\" (UniqueName: \"kubernetes.io/projected/1d618973-0b94-496b-b7eb-d06b37ec6506-kube-api-access-qs5mp\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021174 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9pvm\" (UniqueName: \"kubernetes.io/projected/3f9fc314-be24-41d6-80bd-e4622cd1804e-kube-api-access-g9pvm\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021197 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-metrics\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021214 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021235 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/234a63eb-cada-437c-9df9-d2b1058ba5d4-metallb-excludel2\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021251 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-metrics-certs\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021277 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs75l\" (UniqueName: \"kubernetes.io/projected/234a63eb-cada-437c-9df9-d2b1058ba5d4-kube-api-access-hs75l\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021294 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d618973-0b94-496b-b7eb-d06b37ec6506-cert\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021320 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-conf\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021339 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-sockets\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.021363 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d618973-0b94-496b-b7eb-d06b37ec6506-metrics-certs\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.022092 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-reloader\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.022625 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-startup\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.022963 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-metrics\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.023288 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-sockets\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.024412 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3f9fc314-be24-41d6-80bd-e4622cd1804e-frr-conf\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.030760 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f9fc314-be24-41d6-80bd-e4622cd1804e-metrics-certs\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.031559 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26b44d11-5310-42a7-bbbc-fc95de8743a2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8w262\" (UID: \"26b44d11-5310-42a7-bbbc-fc95de8743a2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.042224 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bkp8\" (UniqueName: \"kubernetes.io/projected/26b44d11-5310-42a7-bbbc-fc95de8743a2-kube-api-access-5bkp8\") pod \"frr-k8s-webhook-server-64bf5d555-8w262\" (UID: \"26b44d11-5310-42a7-bbbc-fc95de8743a2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.042456 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9pvm\" (UniqueName: \"kubernetes.io/projected/3f9fc314-be24-41d6-80bd-e4622cd1804e-kube-api-access-g9pvm\") pod \"frr-k8s-6nskv\" (UID: \"3f9fc314-be24-41d6-80bd-e4622cd1804e\") " pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122285 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d618973-0b94-496b-b7eb-d06b37ec6506-metrics-certs\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122435 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs5mp\" (UniqueName: \"kubernetes.io/projected/1d618973-0b94-496b-b7eb-d06b37ec6506-kube-api-access-qs5mp\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122570 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/234a63eb-cada-437c-9df9-d2b1058ba5d4-metallb-excludel2\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122617 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-metrics-certs\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122667 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs75l\" (UniqueName: \"kubernetes.io/projected/234a63eb-cada-437c-9df9-d2b1058ba5d4-kube-api-access-hs75l\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.122704 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d618973-0b94-496b-b7eb-d06b37ec6506-cert\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: E1007 12:36:13.122752 4702 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 12:36:13 crc kubenswrapper[4702]: E1007 12:36:13.123389 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist podName:234a63eb-cada-437c-9df9-d2b1058ba5d4 nodeName:}" failed. No retries permitted until 2025-10-07 12:36:13.623352708 +0000 UTC m=+729.799088854 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist") pod "speaker-7vt7t" (UID: "234a63eb-cada-437c-9df9-d2b1058ba5d4") : secret "metallb-memberlist" not found Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.123764 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/234a63eb-cada-437c-9df9-d2b1058ba5d4-metallb-excludel2\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.125380 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.126154 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d618973-0b94-496b-b7eb-d06b37ec6506-metrics-certs\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.126348 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-metrics-certs\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.130486 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.137779 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d618973-0b94-496b-b7eb-d06b37ec6506-cert\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.140355 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs75l\" (UniqueName: \"kubernetes.io/projected/234a63eb-cada-437c-9df9-d2b1058ba5d4-kube-api-access-hs75l\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.150995 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.156456 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs5mp\" (UniqueName: \"kubernetes.io/projected/1d618973-0b94-496b-b7eb-d06b37ec6506-kube-api-access-qs5mp\") pod \"controller-68d546b9d8-bwfk4\" (UID: \"1d618973-0b94-496b-b7eb-d06b37ec6506\") " pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.244411 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.425205 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"81f9ba88f76a60d1195eaa9219e6b43953af59dd866e0f264c8d1e53b7fe9a5c"} Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.478224 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8w262"] Oct 07 12:36:13 crc kubenswrapper[4702]: W1007 12:36:13.494029 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26b44d11_5310_42a7_bbbc_fc95de8743a2.slice/crio-bb78ded745f3197de0eb35a8b3c2c7d9e8e08c1b8c95858d2363c503a7cbd511 WatchSource:0}: Error finding container bb78ded745f3197de0eb35a8b3c2c7d9e8e08c1b8c95858d2363c503a7cbd511: Status 404 returned error can't find the container with id bb78ded745f3197de0eb35a8b3c2c7d9e8e08c1b8c95858d2363c503a7cbd511 Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.574683 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-bwfk4"] Oct 07 12:36:13 crc kubenswrapper[4702]: W1007 12:36:13.590001 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d618973_0b94_496b_b7eb_d06b37ec6506.slice/crio-fe45d3d9a1b06111be1d119661a1286714d119bcab07f0de4ecc82c36a60d8e9 WatchSource:0}: Error finding container fe45d3d9a1b06111be1d119661a1286714d119bcab07f0de4ecc82c36a60d8e9: Status 404 returned error can't find the container with id fe45d3d9a1b06111be1d119661a1286714d119bcab07f0de4ecc82c36a60d8e9 Oct 07 12:36:13 crc kubenswrapper[4702]: I1007 12:36:13.632854 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:13 crc kubenswrapper[4702]: E1007 12:36:13.633038 4702 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 12:36:13 crc kubenswrapper[4702]: E1007 12:36:13.633087 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist podName:234a63eb-cada-437c-9df9-d2b1058ba5d4 nodeName:}" failed. No retries permitted until 2025-10-07 12:36:14.633068251 +0000 UTC m=+730.808804367 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist") pod "speaker-7vt7t" (UID: "234a63eb-cada-437c-9df9-d2b1058ba5d4") : secret "metallb-memberlist" not found Oct 07 12:36:14 crc kubenswrapper[4702]: I1007 12:36:14.436090 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-bwfk4" event={"ID":"1d618973-0b94-496b-b7eb-d06b37ec6506","Type":"ContainerStarted","Data":"0c74dee86d59a764be70f61302164a3ab519dd50398234d3be9086ea8b5b345c"} Oct 07 12:36:14 crc kubenswrapper[4702]: I1007 12:36:14.436636 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-bwfk4" event={"ID":"1d618973-0b94-496b-b7eb-d06b37ec6506","Type":"ContainerStarted","Data":"fe45d3d9a1b06111be1d119661a1286714d119bcab07f0de4ecc82c36a60d8e9"} Oct 07 12:36:14 crc kubenswrapper[4702]: I1007 12:36:14.440698 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" event={"ID":"26b44d11-5310-42a7-bbbc-fc95de8743a2","Type":"ContainerStarted","Data":"bb78ded745f3197de0eb35a8b3c2c7d9e8e08c1b8c95858d2363c503a7cbd511"} Oct 07 12:36:14 crc kubenswrapper[4702]: I1007 12:36:14.647264 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:14 crc kubenswrapper[4702]: I1007 12:36:14.675109 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/234a63eb-cada-437c-9df9-d2b1058ba5d4-memberlist\") pod \"speaker-7vt7t\" (UID: \"234a63eb-cada-437c-9df9-d2b1058ba5d4\") " pod="metallb-system/speaker-7vt7t" Oct 07 12:36:14 crc kubenswrapper[4702]: I1007 12:36:14.727739 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7vt7t" Oct 07 12:36:15 crc kubenswrapper[4702]: I1007 12:36:15.447150 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7vt7t" event={"ID":"234a63eb-cada-437c-9df9-d2b1058ba5d4","Type":"ContainerStarted","Data":"c8ebd52112fbefba97b5bff0799f33d3d75985d5f9a14b015b47a8fde13e6cf1"} Oct 07 12:36:15 crc kubenswrapper[4702]: I1007 12:36:15.447222 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7vt7t" event={"ID":"234a63eb-cada-437c-9df9-d2b1058ba5d4","Type":"ContainerStarted","Data":"d62d16063b8a3c6ac4ebb24446254058d26c465baa5ff370051c9a6e597cde77"} Oct 07 12:36:18 crc kubenswrapper[4702]: I1007 12:36:18.483215 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7vt7t" event={"ID":"234a63eb-cada-437c-9df9-d2b1058ba5d4","Type":"ContainerStarted","Data":"3c694a20c5eb7e67208b7b4a12bd6a4109041b698e5b5c188c34041ad1624f50"} Oct 07 12:36:18 crc kubenswrapper[4702]: I1007 12:36:18.483619 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7vt7t" Oct 07 12:36:18 crc kubenswrapper[4702]: I1007 12:36:18.485014 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-bwfk4" event={"ID":"1d618973-0b94-496b-b7eb-d06b37ec6506","Type":"ContainerStarted","Data":"fc4552e69fe5bdfb8f84e1f768886c33f19bbc4d89c152f81517ee51c0dcd804"} Oct 07 12:36:18 crc kubenswrapper[4702]: I1007 12:36:18.485209 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:18 crc kubenswrapper[4702]: I1007 12:36:18.508089 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7vt7t" podStartSLOduration=3.717530028 podStartE2EDuration="6.508065395s" podCreationTimestamp="2025-10-07 12:36:12 +0000 UTC" firstStartedPulling="2025-10-07 12:36:15.091349631 +0000 UTC m=+731.267085747" lastFinishedPulling="2025-10-07 12:36:17.881884978 +0000 UTC m=+734.057621114" observedRunningTime="2025-10-07 12:36:18.505039427 +0000 UTC m=+734.680775543" watchObservedRunningTime="2025-10-07 12:36:18.508065395 +0000 UTC m=+734.683801511" Oct 07 12:36:21 crc kubenswrapper[4702]: I1007 12:36:21.506933 4702 generic.go:334] "Generic (PLEG): container finished" podID="3f9fc314-be24-41d6-80bd-e4622cd1804e" containerID="ca3b09d94534f8bd7f2bb001e4eae18870cdce9d3b8e5eb5dca07a1bb58b70d6" exitCode=0 Oct 07 12:36:21 crc kubenswrapper[4702]: I1007 12:36:21.507050 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerDied","Data":"ca3b09d94534f8bd7f2bb001e4eae18870cdce9d3b8e5eb5dca07a1bb58b70d6"} Oct 07 12:36:21 crc kubenswrapper[4702]: I1007 12:36:21.511402 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" event={"ID":"26b44d11-5310-42a7-bbbc-fc95de8743a2","Type":"ContainerStarted","Data":"b89ea9462232773fceb837c759475250ba9636cada7b4f7600e76beeb30bd798"} Oct 07 12:36:21 crc kubenswrapper[4702]: I1007 12:36:21.511700 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:21 crc kubenswrapper[4702]: I1007 12:36:21.547786 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-bwfk4" podStartSLOduration=5.38500023 podStartE2EDuration="9.547749143s" podCreationTimestamp="2025-10-07 12:36:12 +0000 UTC" firstStartedPulling="2025-10-07 12:36:13.716641962 +0000 UTC m=+729.892378118" lastFinishedPulling="2025-10-07 12:36:17.879390915 +0000 UTC m=+734.055127031" observedRunningTime="2025-10-07 12:36:18.52227049 +0000 UTC m=+734.698006606" watchObservedRunningTime="2025-10-07 12:36:21.547749143 +0000 UTC m=+737.723485299" Oct 07 12:36:21 crc kubenswrapper[4702]: I1007 12:36:21.566528 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" podStartSLOduration=2.092429113 podStartE2EDuration="9.56649767s" podCreationTimestamp="2025-10-07 12:36:12 +0000 UTC" firstStartedPulling="2025-10-07 12:36:13.496616734 +0000 UTC m=+729.672352850" lastFinishedPulling="2025-10-07 12:36:20.970685271 +0000 UTC m=+737.146421407" observedRunningTime="2025-10-07 12:36:21.561780003 +0000 UTC m=+737.737516139" watchObservedRunningTime="2025-10-07 12:36:21.56649767 +0000 UTC m=+737.742233826" Oct 07 12:36:22 crc kubenswrapper[4702]: I1007 12:36:22.522156 4702 generic.go:334] "Generic (PLEG): container finished" podID="3f9fc314-be24-41d6-80bd-e4622cd1804e" containerID="b2eb81b3ba2d44ba656c5eb29def7daf2ebf7675ac2da18fe862d54a1a173ff2" exitCode=0 Oct 07 12:36:22 crc kubenswrapper[4702]: I1007 12:36:22.522269 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerDied","Data":"b2eb81b3ba2d44ba656c5eb29def7daf2ebf7675ac2da18fe862d54a1a173ff2"} Oct 07 12:36:23 crc kubenswrapper[4702]: I1007 12:36:23.249089 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-bwfk4" Oct 07 12:36:23 crc kubenswrapper[4702]: I1007 12:36:23.532869 4702 generic.go:334] "Generic (PLEG): container finished" podID="3f9fc314-be24-41d6-80bd-e4622cd1804e" containerID="683df16bbd2d1b7020b35536bee54732c756c878f2fda22b6c03d158f577f856" exitCode=0 Oct 07 12:36:23 crc kubenswrapper[4702]: I1007 12:36:23.532973 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerDied","Data":"683df16bbd2d1b7020b35536bee54732c756c878f2fda22b6c03d158f577f856"} Oct 07 12:36:24 crc kubenswrapper[4702]: I1007 12:36:24.548584 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"b66a29f996411485e93acffdf2bea408c9d42da7efbda62dcf188aa381112a14"} Oct 07 12:36:24 crc kubenswrapper[4702]: I1007 12:36:24.549125 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"7b92cc122c01b0f88faed75045a6fe603e1467c614587a374b10643edf077efd"} Oct 07 12:36:24 crc kubenswrapper[4702]: I1007 12:36:24.549148 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"8a240f874dbcce3a9eae58b5851a1609b76fbce6f593327b869524df6987c8c1"} Oct 07 12:36:24 crc kubenswrapper[4702]: I1007 12:36:24.549166 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"1da73d081577d03a86f863be578565839693d2fc35425f115b22a72a02b2f5af"} Oct 07 12:36:24 crc kubenswrapper[4702]: I1007 12:36:24.549184 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"a10af0a5eed2bb9958fb83cee69408305e6caabf7728de60191a0d4443c353de"} Oct 07 12:36:25 crc kubenswrapper[4702]: I1007 12:36:25.572576 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6nskv" event={"ID":"3f9fc314-be24-41d6-80bd-e4622cd1804e","Type":"ContainerStarted","Data":"8d9aa7a4aa5637586cf4115b4255227b291b03c0f285b7f1d5ec2f81eccac8c0"} Oct 07 12:36:25 crc kubenswrapper[4702]: I1007 12:36:25.573129 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:25 crc kubenswrapper[4702]: I1007 12:36:25.607505 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6nskv" podStartSLOduration=5.993689806 podStartE2EDuration="13.607462075s" podCreationTimestamp="2025-10-07 12:36:12 +0000 UTC" firstStartedPulling="2025-10-07 12:36:13.384087606 +0000 UTC m=+729.559823722" lastFinishedPulling="2025-10-07 12:36:20.997859865 +0000 UTC m=+737.173595991" observedRunningTime="2025-10-07 12:36:25.605742035 +0000 UTC m=+741.781478161" watchObservedRunningTime="2025-10-07 12:36:25.607462075 +0000 UTC m=+741.783198211" Oct 07 12:36:28 crc kubenswrapper[4702]: I1007 12:36:28.152331 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:28 crc kubenswrapper[4702]: I1007 12:36:28.199143 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.050463 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9lq9"] Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.050790 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerName="controller-manager" containerID="cri-o://071fdaf3c50efd2a7976aab5f915496ca6540743d6a707be4cc7d0cb5c4ec926" gracePeriod=30 Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.129750 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj"] Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.130031 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" podUID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" containerName="route-controller-manager" containerID="cri-o://ad6a88b5458481337f8f8fee3e18c5f6945f7f535564caa59f2db015f0cd0abc" gracePeriod=30 Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.604281 4702 generic.go:334] "Generic (PLEG): container finished" podID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerID="071fdaf3c50efd2a7976aab5f915496ca6540743d6a707be4cc7d0cb5c4ec926" exitCode=0 Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.604383 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" event={"ID":"a3ed3536-f707-47a5-b3df-3ec72bec773b","Type":"ContainerDied","Data":"071fdaf3c50efd2a7976aab5f915496ca6540743d6a707be4cc7d0cb5c4ec926"} Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.606148 4702 generic.go:334] "Generic (PLEG): container finished" podID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" containerID="ad6a88b5458481337f8f8fee3e18c5f6945f7f535564caa59f2db015f0cd0abc" exitCode=0 Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.606202 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" event={"ID":"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c","Type":"ContainerDied","Data":"ad6a88b5458481337f8f8fee3e18c5f6945f7f535564caa59f2db015f0cd0abc"} Oct 07 12:36:29 crc kubenswrapper[4702]: I1007 12:36:29.969337 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.014845 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.018403 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-proxy-ca-bundles\") pod \"a3ed3536-f707-47a5-b3df-3ec72bec773b\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.018477 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnjpg\" (UniqueName: \"kubernetes.io/projected/a3ed3536-f707-47a5-b3df-3ec72bec773b-kube-api-access-vnjpg\") pod \"a3ed3536-f707-47a5-b3df-3ec72bec773b\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.018510 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3ed3536-f707-47a5-b3df-3ec72bec773b-serving-cert\") pod \"a3ed3536-f707-47a5-b3df-3ec72bec773b\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.018618 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-client-ca\") pod \"a3ed3536-f707-47a5-b3df-3ec72bec773b\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.018655 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-config\") pod \"a3ed3536-f707-47a5-b3df-3ec72bec773b\" (UID: \"a3ed3536-f707-47a5-b3df-3ec72bec773b\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.019273 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a3ed3536-f707-47a5-b3df-3ec72bec773b" (UID: "a3ed3536-f707-47a5-b3df-3ec72bec773b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.019422 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-config" (OuterVolumeSpecName: "config") pod "a3ed3536-f707-47a5-b3df-3ec72bec773b" (UID: "a3ed3536-f707-47a5-b3df-3ec72bec773b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.019518 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-client-ca" (OuterVolumeSpecName: "client-ca") pod "a3ed3536-f707-47a5-b3df-3ec72bec773b" (UID: "a3ed3536-f707-47a5-b3df-3ec72bec773b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.032039 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3ed3536-f707-47a5-b3df-3ec72bec773b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a3ed3536-f707-47a5-b3df-3ec72bec773b" (UID: "a3ed3536-f707-47a5-b3df-3ec72bec773b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.037519 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3ed3536-f707-47a5-b3df-3ec72bec773b-kube-api-access-vnjpg" (OuterVolumeSpecName: "kube-api-access-vnjpg") pod "a3ed3536-f707-47a5-b3df-3ec72bec773b" (UID: "a3ed3536-f707-47a5-b3df-3ec72bec773b"). InnerVolumeSpecName "kube-api-access-vnjpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.114562 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8d757454f-wfgf5"] Oct 07 12:36:30 crc kubenswrapper[4702]: E1007 12:36:30.114869 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerName="controller-manager" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.114912 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerName="controller-manager" Oct 07 12:36:30 crc kubenswrapper[4702]: E1007 12:36:30.114925 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" containerName="route-controller-manager" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.114934 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" containerName="route-controller-manager" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.115067 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" containerName="controller-manager" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.115086 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" containerName="route-controller-manager" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.115744 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.119959 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-config\") pod \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120096 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqpk8\" (UniqueName: \"kubernetes.io/projected/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-kube-api-access-lqpk8\") pod \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120120 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-serving-cert\") pod \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120149 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-client-ca\") pod \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\" (UID: \"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c\") " Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120627 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120657 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnjpg\" (UniqueName: \"kubernetes.io/projected/a3ed3536-f707-47a5-b3df-3ec72bec773b-kube-api-access-vnjpg\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120674 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3ed3536-f707-47a5-b3df-3ec72bec773b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120688 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.120699 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ed3536-f707-47a5-b3df-3ec72bec773b-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.121833 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-client-ca" (OuterVolumeSpecName: "client-ca") pod "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" (UID: "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.123694 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-config" (OuterVolumeSpecName: "config") pod "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" (UID: "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.130464 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" (UID: "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.133858 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-kube-api-access-lqpk8" (OuterVolumeSpecName: "kube-api-access-lqpk8") pod "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" (UID: "3f9e5634-520a-4fc6-9bd2-623fe4f4f59c"). InnerVolumeSpecName "kube-api-access-lqpk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.136283 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.137246 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.175995 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.179060 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8d757454f-wfgf5"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.222886 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-proxy-ca-bundles\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.222974 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-serving-cert\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223016 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-client-ca\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223049 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-config\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223077 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8462\" (UniqueName: \"kubernetes.io/projected/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-kube-api-access-t8462\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223099 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-serving-cert\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223145 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdctx\" (UniqueName: \"kubernetes.io/projected/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-kube-api-access-wdctx\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223184 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-config\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223218 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-client-ca\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223282 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqpk8\" (UniqueName: \"kubernetes.io/projected/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-kube-api-access-lqpk8\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223295 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223306 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.223318 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.324823 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdctx\" (UniqueName: \"kubernetes.io/projected/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-kube-api-access-wdctx\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.324922 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-config\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.324969 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-client-ca\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.325003 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-proxy-ca-bundles\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.325025 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-serving-cert\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.325049 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-client-ca\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.325071 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-config\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.325098 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8462\" (UniqueName: \"kubernetes.io/projected/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-kube-api-access-t8462\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.325121 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-serving-cert\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.326668 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-client-ca\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.326694 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-client-ca\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.326861 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-config\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.327051 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-proxy-ca-bundles\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.327494 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-config\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.329735 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-serving-cert\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.331236 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-serving-cert\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.347914 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdctx\" (UniqueName: \"kubernetes.io/projected/ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8-kube-api-access-wdctx\") pod \"route-controller-manager-6b689d5bf9-6fsp4\" (UID: \"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8\") " pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.353679 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8462\" (UniqueName: \"kubernetes.io/projected/a8feeb0c-ff15-4144-988b-9e4a9b3feb5b-kube-api-access-t8462\") pod \"controller-manager-8d757454f-wfgf5\" (UID: \"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b\") " pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.443818 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.456140 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.656688 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.661333 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9lq9" event={"ID":"a3ed3536-f707-47a5-b3df-3ec72bec773b","Type":"ContainerDied","Data":"4ae891340b065f1d228b2317edd9199de9ff971494bb9ad1770418735e63c653"} Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.661373 4702 scope.go:117] "RemoveContainer" containerID="071fdaf3c50efd2a7976aab5f915496ca6540743d6a707be4cc7d0cb5c4ec926" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.676673 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.676378 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj" event={"ID":"3f9e5634-520a-4fc6-9bd2-623fe4f4f59c","Type":"ContainerDied","Data":"a9949c4c1fcc8c61d39103a785a8ebd17877947e92ecfd1740377b683709f359"} Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.699677 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9lq9"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.702111 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9lq9"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.716210 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.725209 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5dwsj"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.725395 4702 scope.go:117] "RemoveContainer" containerID="ad6a88b5458481337f8f8fee3e18c5f6945f7f535564caa59f2db015f0cd0abc" Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.792314 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4"] Oct 07 12:36:30 crc kubenswrapper[4702]: I1007 12:36:30.836512 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8d757454f-wfgf5"] Oct 07 12:36:30 crc kubenswrapper[4702]: W1007 12:36:30.841208 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8feeb0c_ff15_4144_988b_9e4a9b3feb5b.slice/crio-be45222c55754a48ef0542b33635ae862720cd98bb041285981b40c51b7842da WatchSource:0}: Error finding container be45222c55754a48ef0542b33635ae862720cd98bb041285981b40c51b7842da: Status 404 returned error can't find the container with id be45222c55754a48ef0542b33635ae862720cd98bb041285981b40c51b7842da Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.686202 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" event={"ID":"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b","Type":"ContainerStarted","Data":"5457eb903ce10a18ab3f185448811967fe5d4a6aba66e89fafeae4c355eae936"} Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.686646 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" event={"ID":"a8feeb0c-ff15-4144-988b-9e4a9b3feb5b","Type":"ContainerStarted","Data":"be45222c55754a48ef0542b33635ae862720cd98bb041285981b40c51b7842da"} Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.686665 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.687884 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" event={"ID":"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8","Type":"ContainerStarted","Data":"f7606c540d676eb7bca4a8c016ccfa8294c05fb71fdadd5906ea98df33e0f130"} Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.687943 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" event={"ID":"ba60a16a-23a3-4fd0-bc92-8bcfbfde59e8","Type":"ContainerStarted","Data":"a0c92417dd12c1ea17ca20d1d0f31cef69419f592f0e3c0de29e451e161e8c29"} Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.688233 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.693483 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.693618 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" Oct 07 12:36:31 crc kubenswrapper[4702]: I1007 12:36:31.711816 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-8d757454f-wfgf5" podStartSLOduration=1.711790019 podStartE2EDuration="1.711790019s" podCreationTimestamp="2025-10-07 12:36:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:36:31.710275145 +0000 UTC m=+747.886011301" watchObservedRunningTime="2025-10-07 12:36:31.711790019 +0000 UTC m=+747.887526135" Oct 07 12:36:32 crc kubenswrapper[4702]: I1007 12:36:32.647789 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f9e5634-520a-4fc6-9bd2-623fe4f4f59c" path="/var/lib/kubelet/pods/3f9e5634-520a-4fc6-9bd2-623fe4f4f59c/volumes" Oct 07 12:36:32 crc kubenswrapper[4702]: I1007 12:36:32.648410 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3ed3536-f707-47a5-b3df-3ec72bec773b" path="/var/lib/kubelet/pods/a3ed3536-f707-47a5-b3df-3ec72bec773b/volumes" Oct 07 12:36:33 crc kubenswrapper[4702]: I1007 12:36:33.139617 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8w262" Oct 07 12:36:33 crc kubenswrapper[4702]: I1007 12:36:33.170207 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b689d5bf9-6fsp4" podStartSLOduration=3.170174923 podStartE2EDuration="3.170174923s" podCreationTimestamp="2025-10-07 12:36:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:36:31.799735879 +0000 UTC m=+747.975471985" watchObservedRunningTime="2025-10-07 12:36:33.170174923 +0000 UTC m=+749.345911049" Oct 07 12:36:33 crc kubenswrapper[4702]: I1007 12:36:33.171652 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6nskv" Oct 07 12:36:34 crc kubenswrapper[4702]: I1007 12:36:34.735065 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7vt7t" Oct 07 12:36:39 crc kubenswrapper[4702]: I1007 12:36:39.262623 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:36:39 crc kubenswrapper[4702]: I1007 12:36:39.264690 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:36:39 crc kubenswrapper[4702]: I1007 12:36:39.265074 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:36:39 crc kubenswrapper[4702]: I1007 12:36:39.266595 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef24fdfca720c04e2e8d0834767736417773f1bf6dd8396fbe53fc5a646c234f"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:36:39 crc kubenswrapper[4702]: I1007 12:36:39.266959 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://ef24fdfca720c04e2e8d0834767736417773f1bf6dd8396fbe53fc5a646c234f" gracePeriod=600 Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.758079 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="ef24fdfca720c04e2e8d0834767736417773f1bf6dd8396fbe53fc5a646c234f" exitCode=0 Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.758275 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"ef24fdfca720c04e2e8d0834767736417773f1bf6dd8396fbe53fc5a646c234f"} Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.758693 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"3ca59d66613e781014a1bca3d21d490470064dd363a5feb6c9d2c05e7748d68e"} Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.758726 4702 scope.go:117] "RemoveContainer" containerID="dbea129310a361471f0e292025f60b439ffa2aae1c5e93aa00ef5e1ce0d3006d" Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.838969 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-xq49x"] Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.839693 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.850876 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-xq49x"] Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.870057 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.871350 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhr4f\" (UniqueName: \"kubernetes.io/projected/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e-kube-api-access-lhr4f\") pod \"mariadb-operator-index-xq49x\" (UID: \"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e\") " pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.872923 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-r6t8k" Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.874580 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 07 12:36:40 crc kubenswrapper[4702]: I1007 12:36:40.972548 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhr4f\" (UniqueName: \"kubernetes.io/projected/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e-kube-api-access-lhr4f\") pod \"mariadb-operator-index-xq49x\" (UID: \"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e\") " pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:41 crc kubenswrapper[4702]: I1007 12:36:41.012878 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhr4f\" (UniqueName: \"kubernetes.io/projected/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e-kube-api-access-lhr4f\") pod \"mariadb-operator-index-xq49x\" (UID: \"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e\") " pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:41 crc kubenswrapper[4702]: I1007 12:36:41.188351 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:41 crc kubenswrapper[4702]: I1007 12:36:41.624199 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-xq49x"] Oct 07 12:36:41 crc kubenswrapper[4702]: W1007 12:36:41.635595 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3fe5fb5_75c5_4e6e_8c45_c0b7121eaf5e.slice/crio-15b71a0dd4c1da7fd16b08138b26d3d4ee809a1bbafb11a01cba90e5d6a3ed48 WatchSource:0}: Error finding container 15b71a0dd4c1da7fd16b08138b26d3d4ee809a1bbafb11a01cba90e5d6a3ed48: Status 404 returned error can't find the container with id 15b71a0dd4c1da7fd16b08138b26d3d4ee809a1bbafb11a01cba90e5d6a3ed48 Oct 07 12:36:41 crc kubenswrapper[4702]: I1007 12:36:41.769089 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-xq49x" event={"ID":"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e","Type":"ContainerStarted","Data":"15b71a0dd4c1da7fd16b08138b26d3d4ee809a1bbafb11a01cba90e5d6a3ed48"} Oct 07 12:36:43 crc kubenswrapper[4702]: I1007 12:36:43.790893 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-xq49x" event={"ID":"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e","Type":"ContainerStarted","Data":"e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1"} Oct 07 12:36:43 crc kubenswrapper[4702]: I1007 12:36:43.811479 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-xq49x" podStartSLOduration=2.893534133 podStartE2EDuration="3.811448214s" podCreationTimestamp="2025-10-07 12:36:40 +0000 UTC" firstStartedPulling="2025-10-07 12:36:41.6380752 +0000 UTC m=+757.813811316" lastFinishedPulling="2025-10-07 12:36:42.555989281 +0000 UTC m=+758.731725397" observedRunningTime="2025-10-07 12:36:43.807035316 +0000 UTC m=+759.982771432" watchObservedRunningTime="2025-10-07 12:36:43.811448214 +0000 UTC m=+759.987184370" Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.009416 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-xq49x"] Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.625425 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-xn9xd"] Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.626179 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.626912 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxlf2\" (UniqueName: \"kubernetes.io/projected/0591047e-35dd-4650-8730-d7795ce53a56-kube-api-access-gxlf2\") pod \"mariadb-operator-index-xn9xd\" (UID: \"0591047e-35dd-4650-8730-d7795ce53a56\") " pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.650533 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-xn9xd"] Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.728760 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxlf2\" (UniqueName: \"kubernetes.io/projected/0591047e-35dd-4650-8730-d7795ce53a56-kube-api-access-gxlf2\") pod \"mariadb-operator-index-xn9xd\" (UID: \"0591047e-35dd-4650-8730-d7795ce53a56\") " pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.761280 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxlf2\" (UniqueName: \"kubernetes.io/projected/0591047e-35dd-4650-8730-d7795ce53a56-kube-api-access-gxlf2\") pod \"mariadb-operator-index-xn9xd\" (UID: \"0591047e-35dd-4650-8730-d7795ce53a56\") " pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:44 crc kubenswrapper[4702]: I1007 12:36:44.942692 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:45 crc kubenswrapper[4702]: I1007 12:36:45.408563 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-xn9xd"] Oct 07 12:36:45 crc kubenswrapper[4702]: I1007 12:36:45.804576 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-xn9xd" event={"ID":"0591047e-35dd-4650-8730-d7795ce53a56","Type":"ContainerStarted","Data":"921f5c1f6741f1add2a72aa43738fcf444667bb55e3541259fa1926ed166e28a"} Oct 07 12:36:45 crc kubenswrapper[4702]: I1007 12:36:45.804767 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-xq49x" podUID="d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" containerName="registry-server" containerID="cri-o://e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1" gracePeriod=2 Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.241588 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.353061 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhr4f\" (UniqueName: \"kubernetes.io/projected/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e-kube-api-access-lhr4f\") pod \"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e\" (UID: \"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e\") " Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.361122 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e-kube-api-access-lhr4f" (OuterVolumeSpecName: "kube-api-access-lhr4f") pod "d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" (UID: "d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e"). InnerVolumeSpecName "kube-api-access-lhr4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.454791 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhr4f\" (UniqueName: \"kubernetes.io/projected/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e-kube-api-access-lhr4f\") on node \"crc\" DevicePath \"\"" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.822131 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-xn9xd" event={"ID":"0591047e-35dd-4650-8730-d7795ce53a56","Type":"ContainerStarted","Data":"e9dcf04daf903c7f18cc9ec315d2ef15eb916f019193d6956b28c5d97494ec25"} Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.827405 4702 generic.go:334] "Generic (PLEG): container finished" podID="d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" containerID="e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1" exitCode=0 Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.827471 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-xq49x" event={"ID":"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e","Type":"ContainerDied","Data":"e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1"} Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.827511 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-xq49x" event={"ID":"d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e","Type":"ContainerDied","Data":"15b71a0dd4c1da7fd16b08138b26d3d4ee809a1bbafb11a01cba90e5d6a3ed48"} Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.827541 4702 scope.go:117] "RemoveContainer" containerID="e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.827703 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-xq49x" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.851789 4702 scope.go:117] "RemoveContainer" containerID="e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.852951 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-xn9xd" podStartSLOduration=2.353439491 podStartE2EDuration="2.852922754s" podCreationTimestamp="2025-10-07 12:36:44 +0000 UTC" firstStartedPulling="2025-10-07 12:36:45.433269564 +0000 UTC m=+761.609005690" lastFinishedPulling="2025-10-07 12:36:45.932752837 +0000 UTC m=+762.108488953" observedRunningTime="2025-10-07 12:36:46.851459032 +0000 UTC m=+763.027195148" watchObservedRunningTime="2025-10-07 12:36:46.852922754 +0000 UTC m=+763.028658890" Oct 07 12:36:46 crc kubenswrapper[4702]: E1007 12:36:46.853996 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1\": container with ID starting with e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1 not found: ID does not exist" containerID="e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.854029 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1"} err="failed to get container status \"e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1\": rpc error: code = NotFound desc = could not find container \"e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1\": container with ID starting with e7129b6f9647b118ed2ab9ed5bfe4335d8667d60875444936e5f1da40117d4e1 not found: ID does not exist" Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.870916 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-xq49x"] Oct 07 12:36:46 crc kubenswrapper[4702]: I1007 12:36:46.875639 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-xq49x"] Oct 07 12:36:48 crc kubenswrapper[4702]: I1007 12:36:48.649477 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" path="/var/lib/kubelet/pods/d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e/volumes" Oct 07 12:36:54 crc kubenswrapper[4702]: I1007 12:36:54.943592 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:54 crc kubenswrapper[4702]: I1007 12:36:54.944581 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:54 crc kubenswrapper[4702]: I1007 12:36:54.982720 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:55 crc kubenswrapper[4702]: I1007 12:36:55.935041 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-xn9xd" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.273454 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t"] Oct 07 12:36:57 crc kubenswrapper[4702]: E1007 12:36:57.273886 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" containerName="registry-server" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.273946 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" containerName="registry-server" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.274198 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fe5fb5-75c5-4e6e-8c45-c0b7121eaf5e" containerName="registry-server" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.275693 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.281408 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrk4d" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.288103 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t"] Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.421508 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv54k\" (UniqueName: \"kubernetes.io/projected/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-kube-api-access-tv54k\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.421577 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-bundle\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.421723 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-util\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.523053 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-util\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.523686 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv54k\" (UniqueName: \"kubernetes.io/projected/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-kube-api-access-tv54k\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.523569 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-util\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.523725 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-bundle\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.524205 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-bundle\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.556171 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv54k\" (UniqueName: \"kubernetes.io/projected/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-kube-api-access-tv54k\") pod \"10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:57 crc kubenswrapper[4702]: I1007 12:36:57.628441 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:36:58 crc kubenswrapper[4702]: I1007 12:36:58.084277 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t"] Oct 07 12:36:58 crc kubenswrapper[4702]: I1007 12:36:58.922536 4702 generic.go:334] "Generic (PLEG): container finished" podID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerID="dc1508d573a1a88b9e3dc076cfebf115b8e8afae779f96285d6d119d1ba90e58" exitCode=0 Oct 07 12:36:58 crc kubenswrapper[4702]: I1007 12:36:58.922595 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" event={"ID":"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e","Type":"ContainerDied","Data":"dc1508d573a1a88b9e3dc076cfebf115b8e8afae779f96285d6d119d1ba90e58"} Oct 07 12:36:58 crc kubenswrapper[4702]: I1007 12:36:58.922628 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" event={"ID":"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e","Type":"ContainerStarted","Data":"f960e4654d230af4a4ecb126258840e17acb6c82fab8b3aba7ac96e0cc94359e"} Oct 07 12:36:59 crc kubenswrapper[4702]: I1007 12:36:59.933734 4702 generic.go:334] "Generic (PLEG): container finished" podID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerID="c58d19ab16ed7f88906463a743a2c4c22321d6fb9bda0906ad860db60db6b6ae" exitCode=0 Oct 07 12:36:59 crc kubenswrapper[4702]: I1007 12:36:59.933833 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" event={"ID":"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e","Type":"ContainerDied","Data":"c58d19ab16ed7f88906463a743a2c4c22321d6fb9bda0906ad860db60db6b6ae"} Oct 07 12:37:00 crc kubenswrapper[4702]: I1007 12:37:00.947134 4702 generic.go:334] "Generic (PLEG): container finished" podID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerID="90fc70a824e3582a820afaae462fdf36ae13811ce74abdedfdb450de78b5f491" exitCode=0 Oct 07 12:37:00 crc kubenswrapper[4702]: I1007 12:37:00.947218 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" event={"ID":"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e","Type":"ContainerDied","Data":"90fc70a824e3582a820afaae462fdf36ae13811ce74abdedfdb450de78b5f491"} Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.392993 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.398105 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv54k\" (UniqueName: \"kubernetes.io/projected/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-kube-api-access-tv54k\") pod \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.398164 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-util\") pod \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.398191 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-bundle\") pod \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\" (UID: \"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e\") " Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.399308 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-bundle" (OuterVolumeSpecName: "bundle") pod "9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" (UID: "9b6ccce5-1320-48c5-b65f-6f0cfd8d771e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.409177 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-kube-api-access-tv54k" (OuterVolumeSpecName: "kube-api-access-tv54k") pod "9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" (UID: "9b6ccce5-1320-48c5-b65f-6f0cfd8d771e"). InnerVolumeSpecName "kube-api-access-tv54k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.415430 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-util" (OuterVolumeSpecName: "util") pod "9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" (UID: "9b6ccce5-1320-48c5-b65f-6f0cfd8d771e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.498817 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.498851 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.498860 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv54k\" (UniqueName: \"kubernetes.io/projected/9b6ccce5-1320-48c5-b65f-6f0cfd8d771e-kube-api-access-tv54k\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.967168 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" event={"ID":"9b6ccce5-1320-48c5-b65f-6f0cfd8d771e","Type":"ContainerDied","Data":"f960e4654d230af4a4ecb126258840e17acb6c82fab8b3aba7ac96e0cc94359e"} Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.967235 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f960e4654d230af4a4ecb126258840e17acb6c82fab8b3aba7ac96e0cc94359e" Oct 07 12:37:02 crc kubenswrapper[4702]: I1007 12:37:02.967291 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.008055 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp"] Oct 07 12:37:11 crc kubenswrapper[4702]: E1007 12:37:11.009029 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="extract" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.009044 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="extract" Oct 07 12:37:11 crc kubenswrapper[4702]: E1007 12:37:11.009053 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="pull" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.009059 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="pull" Oct 07 12:37:11 crc kubenswrapper[4702]: E1007 12:37:11.009082 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="util" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.009089 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="util" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.009228 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b6ccce5-1320-48c5-b65f-6f0cfd8d771e" containerName="extract" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.009932 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.016035 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.016072 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.019234 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tpmf9" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.029671 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-apiservice-cert\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.029737 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49bjl\" (UniqueName: \"kubernetes.io/projected/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-kube-api-access-49bjl\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.029887 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-webhook-cert\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.037849 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp"] Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.130773 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49bjl\" (UniqueName: \"kubernetes.io/projected/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-kube-api-access-49bjl\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.130864 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-webhook-cert\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.130935 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-apiservice-cert\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.140710 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-apiservice-cert\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.143386 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-webhook-cert\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.149635 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49bjl\" (UniqueName: \"kubernetes.io/projected/126878fb-9a6e-4cf2-9bd5-cfcfab6888c6-kube-api-access-49bjl\") pod \"mariadb-operator-controller-manager-79c58767fd-n8qwp\" (UID: \"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.329766 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:11 crc kubenswrapper[4702]: I1007 12:37:11.753140 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp"] Oct 07 12:37:12 crc kubenswrapper[4702]: I1007 12:37:12.027972 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" event={"ID":"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6","Type":"ContainerStarted","Data":"3a9baa89bc4f98bd392f6dd8d8502f4c29011238006fe287703aec18227748f4"} Oct 07 12:37:16 crc kubenswrapper[4702]: I1007 12:37:16.075963 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" event={"ID":"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6","Type":"ContainerStarted","Data":"c74842ae42c1cf35ba17a3228d1af6af6a7377581d15d9eaceaa02c0b3b4752d"} Oct 07 12:37:19 crc kubenswrapper[4702]: I1007 12:37:19.097634 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" event={"ID":"126878fb-9a6e-4cf2-9bd5-cfcfab6888c6","Type":"ContainerStarted","Data":"6ee6a35857174ce6a0c3eb7078b88f7caffed185862f62b1a895b68a3d882a80"} Oct 07 12:37:19 crc kubenswrapper[4702]: I1007 12:37:19.098449 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:20 crc kubenswrapper[4702]: I1007 12:37:20.108780 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" Oct 07 12:37:20 crc kubenswrapper[4702]: I1007 12:37:20.130859 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c58767fd-n8qwp" podStartSLOduration=3.7193966979999997 podStartE2EDuration="10.130829574s" podCreationTimestamp="2025-10-07 12:37:10 +0000 UTC" firstStartedPulling="2025-10-07 12:37:11.768174967 +0000 UTC m=+787.943911073" lastFinishedPulling="2025-10-07 12:37:18.179607833 +0000 UTC m=+794.355343949" observedRunningTime="2025-10-07 12:37:19.134083387 +0000 UTC m=+795.309819543" watchObservedRunningTime="2025-10-07 12:37:20.130829574 +0000 UTC m=+796.306565720" Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.341535 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-2zrr4"] Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.342946 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.358755 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk47s\" (UniqueName: \"kubernetes.io/projected/da437e15-80b4-4b13-8c1e-a117c6084274-kube-api-access-fk47s\") pod \"infra-operator-index-2zrr4\" (UID: \"da437e15-80b4-4b13-8c1e-a117c6084274\") " pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.360559 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-7q5kn" Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.372459 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-2zrr4"] Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.461041 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk47s\" (UniqueName: \"kubernetes.io/projected/da437e15-80b4-4b13-8c1e-a117c6084274-kube-api-access-fk47s\") pod \"infra-operator-index-2zrr4\" (UID: \"da437e15-80b4-4b13-8c1e-a117c6084274\") " pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.484263 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk47s\" (UniqueName: \"kubernetes.io/projected/da437e15-80b4-4b13-8c1e-a117c6084274-kube-api-access-fk47s\") pod \"infra-operator-index-2zrr4\" (UID: \"da437e15-80b4-4b13-8c1e-a117c6084274\") " pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:26 crc kubenswrapper[4702]: I1007 12:37:26.668661 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:27 crc kubenswrapper[4702]: I1007 12:37:27.111006 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-2zrr4"] Oct 07 12:37:27 crc kubenswrapper[4702]: W1007 12:37:27.118228 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda437e15_80b4_4b13_8c1e_a117c6084274.slice/crio-4e4ad849bd44bcfca1f5ba3e4686bbe05029d0d3236e92a4b604b1366ed7669b WatchSource:0}: Error finding container 4e4ad849bd44bcfca1f5ba3e4686bbe05029d0d3236e92a4b604b1366ed7669b: Status 404 returned error can't find the container with id 4e4ad849bd44bcfca1f5ba3e4686bbe05029d0d3236e92a4b604b1366ed7669b Oct 07 12:37:27 crc kubenswrapper[4702]: I1007 12:37:27.143510 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2zrr4" event={"ID":"da437e15-80b4-4b13-8c1e-a117c6084274","Type":"ContainerStarted","Data":"4e4ad849bd44bcfca1f5ba3e4686bbe05029d0d3236e92a4b604b1366ed7669b"} Oct 07 12:37:30 crc kubenswrapper[4702]: I1007 12:37:30.169387 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2zrr4" event={"ID":"da437e15-80b4-4b13-8c1e-a117c6084274","Type":"ContainerStarted","Data":"513c62f2b8afbff4df95d320b93dd6fec88bf411271b7f2fe38d5a4c65121bf1"} Oct 07 12:37:31 crc kubenswrapper[4702]: I1007 12:37:31.193006 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-2zrr4" podStartSLOduration=2.497129137 podStartE2EDuration="5.19297839s" podCreationTimestamp="2025-10-07 12:37:26 +0000 UTC" firstStartedPulling="2025-10-07 12:37:27.121380276 +0000 UTC m=+803.297116392" lastFinishedPulling="2025-10-07 12:37:29.817229509 +0000 UTC m=+805.992965645" observedRunningTime="2025-10-07 12:37:31.192499857 +0000 UTC m=+807.368235973" watchObservedRunningTime="2025-10-07 12:37:31.19297839 +0000 UTC m=+807.368714516" Oct 07 12:37:35 crc kubenswrapper[4702]: I1007 12:37:35.746040 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d8twc"] Oct 07 12:37:35 crc kubenswrapper[4702]: I1007 12:37:35.748281 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:35 crc kubenswrapper[4702]: I1007 12:37:35.774528 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8twc"] Oct 07 12:37:35 crc kubenswrapper[4702]: I1007 12:37:35.915651 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfvb\" (UniqueName: \"kubernetes.io/projected/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-kube-api-access-8dfvb\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:35 crc kubenswrapper[4702]: I1007 12:37:35.915846 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-utilities\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:35 crc kubenswrapper[4702]: I1007 12:37:35.915943 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-catalog-content\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.017033 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfvb\" (UniqueName: \"kubernetes.io/projected/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-kube-api-access-8dfvb\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.017108 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-utilities\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.017137 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-catalog-content\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.017663 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-catalog-content\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.017694 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-utilities\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.037620 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfvb\" (UniqueName: \"kubernetes.io/projected/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-kube-api-access-8dfvb\") pod \"redhat-marketplace-d8twc\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.119447 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.587348 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8twc"] Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.669038 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.669104 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:36 crc kubenswrapper[4702]: I1007 12:37:36.731008 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:37 crc kubenswrapper[4702]: I1007 12:37:37.217889 4702 generic.go:334] "Generic (PLEG): container finished" podID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerID="05421109b812a4c37941cea9cdf6ca10c50bac2230924ddaeedf0196465c2f2e" exitCode=0 Oct 07 12:37:37 crc kubenswrapper[4702]: I1007 12:37:37.218037 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8twc" event={"ID":"e91d5956-74c0-4cf4-ac24-29b3a840a7d7","Type":"ContainerDied","Data":"05421109b812a4c37941cea9cdf6ca10c50bac2230924ddaeedf0196465c2f2e"} Oct 07 12:37:37 crc kubenswrapper[4702]: I1007 12:37:37.218106 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8twc" event={"ID":"e91d5956-74c0-4cf4-ac24-29b3a840a7d7","Type":"ContainerStarted","Data":"5ca2acee1a2fee7709b4c418aa109d2031fda374629c68a63614a724e51f5bf4"} Oct 07 12:37:37 crc kubenswrapper[4702]: I1007 12:37:37.249474 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-2zrr4" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.234440 4702 generic.go:334] "Generic (PLEG): container finished" podID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerID="65bf2fff41acc43d9566e0aaa8b3f71bd430da55a7cc3e1e4983aaa60927dcf6" exitCode=0 Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.234562 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8twc" event={"ID":"e91d5956-74c0-4cf4-ac24-29b3a840a7d7","Type":"ContainerDied","Data":"65bf2fff41acc43d9566e0aaa8b3f71bd430da55a7cc3e1e4983aaa60927dcf6"} Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.379649 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2"] Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.381021 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.386053 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrk4d" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.396079 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2"] Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.478688 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddd4p\" (UniqueName: \"kubernetes.io/projected/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-kube-api-access-ddd4p\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.478841 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-bundle\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.478890 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-util\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.579888 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddd4p\" (UniqueName: \"kubernetes.io/projected/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-kube-api-access-ddd4p\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.580006 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-bundle\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.580041 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-util\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.580743 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-util\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.581168 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-bundle\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.615079 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddd4p\" (UniqueName: \"kubernetes.io/projected/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-kube-api-access-ddd4p\") pod \"4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:39 crc kubenswrapper[4702]: I1007 12:37:39.697683 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:40 crc kubenswrapper[4702]: I1007 12:37:40.199011 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2"] Oct 07 12:37:40 crc kubenswrapper[4702]: I1007 12:37:40.243464 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" event={"ID":"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8","Type":"ContainerStarted","Data":"c64757e8a7971d571ca6f369ed4aa42d5598fc62d468705db15dac2de6ebf426"} Oct 07 12:37:40 crc kubenswrapper[4702]: I1007 12:37:40.248253 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8twc" event={"ID":"e91d5956-74c0-4cf4-ac24-29b3a840a7d7","Type":"ContainerStarted","Data":"256d7c72261918a8425218e07b804dcb67ed73a4fe602f8ef7c11e019828c016"} Oct 07 12:37:40 crc kubenswrapper[4702]: I1007 12:37:40.274741 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d8twc" podStartSLOduration=2.575632405 podStartE2EDuration="5.274718931s" podCreationTimestamp="2025-10-07 12:37:35 +0000 UTC" firstStartedPulling="2025-10-07 12:37:37.21979984 +0000 UTC m=+813.395535956" lastFinishedPulling="2025-10-07 12:37:39.918886366 +0000 UTC m=+816.094622482" observedRunningTime="2025-10-07 12:37:40.273437024 +0000 UTC m=+816.449173140" watchObservedRunningTime="2025-10-07 12:37:40.274718931 +0000 UTC m=+816.450455037" Oct 07 12:37:41 crc kubenswrapper[4702]: I1007 12:37:41.155949 4702 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 12:37:41 crc kubenswrapper[4702]: I1007 12:37:41.255179 4702 generic.go:334] "Generic (PLEG): container finished" podID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerID="14b5ae444a489c71a0e85c5853cfe1cf03c842cd9f63252815845242992b9104" exitCode=0 Oct 07 12:37:41 crc kubenswrapper[4702]: I1007 12:37:41.256364 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" event={"ID":"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8","Type":"ContainerDied","Data":"14b5ae444a489c71a0e85c5853cfe1cf03c842cd9f63252815845242992b9104"} Oct 07 12:37:42 crc kubenswrapper[4702]: I1007 12:37:42.266555 4702 generic.go:334] "Generic (PLEG): container finished" podID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerID="e95ec23dc9a2061ac3daf3d27f4a0d382a708b8b13b1244badbefb898413f3e4" exitCode=0 Oct 07 12:37:42 crc kubenswrapper[4702]: I1007 12:37:42.266649 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" event={"ID":"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8","Type":"ContainerDied","Data":"e95ec23dc9a2061ac3daf3d27f4a0d382a708b8b13b1244badbefb898413f3e4"} Oct 07 12:37:43 crc kubenswrapper[4702]: I1007 12:37:43.277221 4702 generic.go:334] "Generic (PLEG): container finished" podID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerID="c3b3b22bb14907160fc86839b4a5d3e218c1b68c8288f32af26ebe321307fedf" exitCode=0 Oct 07 12:37:43 crc kubenswrapper[4702]: I1007 12:37:43.277500 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" event={"ID":"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8","Type":"ContainerDied","Data":"c3b3b22bb14907160fc86839b4a5d3e218c1b68c8288f32af26ebe321307fedf"} Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.569515 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.660693 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddd4p\" (UniqueName: \"kubernetes.io/projected/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-kube-api-access-ddd4p\") pod \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.660871 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-bundle\") pod \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.660985 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-util\") pod \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\" (UID: \"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8\") " Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.662753 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-bundle" (OuterVolumeSpecName: "bundle") pod "cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" (UID: "cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.674124 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-kube-api-access-ddd4p" (OuterVolumeSpecName: "kube-api-access-ddd4p") pod "cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" (UID: "cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8"). InnerVolumeSpecName "kube-api-access-ddd4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.687215 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-util" (OuterVolumeSpecName: "util") pod "cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" (UID: "cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.762622 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddd4p\" (UniqueName: \"kubernetes.io/projected/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-kube-api-access-ddd4p\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.762677 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:44 crc kubenswrapper[4702]: I1007 12:37:44.762696 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:45 crc kubenswrapper[4702]: I1007 12:37:45.293815 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" event={"ID":"cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8","Type":"ContainerDied","Data":"c64757e8a7971d571ca6f369ed4aa42d5598fc62d468705db15dac2de6ebf426"} Oct 07 12:37:45 crc kubenswrapper[4702]: I1007 12:37:45.294289 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c64757e8a7971d571ca6f369ed4aa42d5598fc62d468705db15dac2de6ebf426" Oct 07 12:37:45 crc kubenswrapper[4702]: I1007 12:37:45.293942 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2" Oct 07 12:37:46 crc kubenswrapper[4702]: I1007 12:37:46.120173 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:46 crc kubenswrapper[4702]: I1007 12:37:46.120228 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:46 crc kubenswrapper[4702]: I1007 12:37:46.188669 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:46 crc kubenswrapper[4702]: I1007 12:37:46.372995 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.335490 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8twc"] Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.336306 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d8twc" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="registry-server" containerID="cri-o://256d7c72261918a8425218e07b804dcb67ed73a4fe602f8ef7c11e019828c016" gracePeriod=2 Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.749743 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kkbfq"] Oct 07 12:37:48 crc kubenswrapper[4702]: E1007 12:37:48.750313 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="extract" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.750361 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="extract" Oct 07 12:37:48 crc kubenswrapper[4702]: E1007 12:37:48.750390 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="pull" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.750411 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="pull" Oct 07 12:37:48 crc kubenswrapper[4702]: E1007 12:37:48.750433 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="util" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.750452 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="util" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.750721 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8" containerName="extract" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.752526 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.760841 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkbfq"] Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.821847 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-utilities\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.821936 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qdzg\" (UniqueName: \"kubernetes.io/projected/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-kube-api-access-7qdzg\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.822183 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-catalog-content\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.923524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-utilities\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.923615 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qdzg\" (UniqueName: \"kubernetes.io/projected/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-kube-api-access-7qdzg\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.923693 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-catalog-content\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.924435 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-catalog-content\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.924633 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-utilities\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:48 crc kubenswrapper[4702]: I1007 12:37:48.955974 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qdzg\" (UniqueName: \"kubernetes.io/projected/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-kube-api-access-7qdzg\") pod \"community-operators-kkbfq\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.148765 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.329365 4702 generic.go:334] "Generic (PLEG): container finished" podID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerID="256d7c72261918a8425218e07b804dcb67ed73a4fe602f8ef7c11e019828c016" exitCode=0 Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.329422 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8twc" event={"ID":"e91d5956-74c0-4cf4-ac24-29b3a840a7d7","Type":"ContainerDied","Data":"256d7c72261918a8425218e07b804dcb67ed73a4fe602f8ef7c11e019828c016"} Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.402555 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.435101 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-catalog-content\") pod \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.435156 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-utilities\") pod \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.435267 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfvb\" (UniqueName: \"kubernetes.io/projected/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-kube-api-access-8dfvb\") pod \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\" (UID: \"e91d5956-74c0-4cf4-ac24-29b3a840a7d7\") " Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.436490 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-utilities" (OuterVolumeSpecName: "utilities") pod "e91d5956-74c0-4cf4-ac24-29b3a840a7d7" (UID: "e91d5956-74c0-4cf4-ac24-29b3a840a7d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.446777 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-kube-api-access-8dfvb" (OuterVolumeSpecName: "kube-api-access-8dfvb") pod "e91d5956-74c0-4cf4-ac24-29b3a840a7d7" (UID: "e91d5956-74c0-4cf4-ac24-29b3a840a7d7"). InnerVolumeSpecName "kube-api-access-8dfvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.450763 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e91d5956-74c0-4cf4-ac24-29b3a840a7d7" (UID: "e91d5956-74c0-4cf4-ac24-29b3a840a7d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.536749 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.536793 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.536806 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfvb\" (UniqueName: \"kubernetes.io/projected/e91d5956-74c0-4cf4-ac24-29b3a840a7d7-kube-api-access-8dfvb\") on node \"crc\" DevicePath \"\"" Oct 07 12:37:49 crc kubenswrapper[4702]: I1007 12:37:49.657415 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkbfq"] Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.337765 4702 generic.go:334] "Generic (PLEG): container finished" podID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerID="e0843d540bdc986381de3039825519bf313c4ac3cac2a53f1e7c5432f2fd737a" exitCode=0 Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.337858 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkbfq" event={"ID":"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2","Type":"ContainerDied","Data":"e0843d540bdc986381de3039825519bf313c4ac3cac2a53f1e7c5432f2fd737a"} Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.337936 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkbfq" event={"ID":"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2","Type":"ContainerStarted","Data":"6efb27a7bdec3f8586530678b32308fd42369c30464434fd8a4bd6afcf9237d5"} Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.341008 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8twc" event={"ID":"e91d5956-74c0-4cf4-ac24-29b3a840a7d7","Type":"ContainerDied","Data":"5ca2acee1a2fee7709b4c418aa109d2031fda374629c68a63614a724e51f5bf4"} Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.341082 4702 scope.go:117] "RemoveContainer" containerID="256d7c72261918a8425218e07b804dcb67ed73a4fe602f8ef7c11e019828c016" Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.341092 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8twc" Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.372330 4702 scope.go:117] "RemoveContainer" containerID="65bf2fff41acc43d9566e0aaa8b3f71bd430da55a7cc3e1e4983aaa60927dcf6" Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.398817 4702 scope.go:117] "RemoveContainer" containerID="05421109b812a4c37941cea9cdf6ca10c50bac2230924ddaeedf0196465c2f2e" Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.404054 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8twc"] Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.414652 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8twc"] Oct 07 12:37:50 crc kubenswrapper[4702]: I1007 12:37:50.647711 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" path="/var/lib/kubelet/pods/e91d5956-74c0-4cf4-ac24-29b3a840a7d7/volumes" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.893275 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p"] Oct 07 12:37:51 crc kubenswrapper[4702]: E1007 12:37:51.894218 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="extract-content" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.894240 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="extract-content" Oct 07 12:37:51 crc kubenswrapper[4702]: E1007 12:37:51.894253 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="registry-server" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.894280 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="registry-server" Oct 07 12:37:51 crc kubenswrapper[4702]: E1007 12:37:51.894295 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="extract-utilities" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.894304 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="extract-utilities" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.894532 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e91d5956-74c0-4cf4-ac24-29b3a840a7d7" containerName="registry-server" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.895399 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.898949 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.903795 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dl5bd" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.919106 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p"] Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.968934 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lc98\" (UniqueName: \"kubernetes.io/projected/beb78a59-5cba-42ca-a7bb-5351e77835b8-kube-api-access-5lc98\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.969028 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/beb78a59-5cba-42ca-a7bb-5351e77835b8-webhook-cert\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:51 crc kubenswrapper[4702]: I1007 12:37:51.969086 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/beb78a59-5cba-42ca-a7bb-5351e77835b8-apiservice-cert\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.070162 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/beb78a59-5cba-42ca-a7bb-5351e77835b8-webhook-cert\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.070213 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/beb78a59-5cba-42ca-a7bb-5351e77835b8-apiservice-cert\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.070269 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lc98\" (UniqueName: \"kubernetes.io/projected/beb78a59-5cba-42ca-a7bb-5351e77835b8-kube-api-access-5lc98\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.077472 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/beb78a59-5cba-42ca-a7bb-5351e77835b8-apiservice-cert\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.079546 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/beb78a59-5cba-42ca-a7bb-5351e77835b8-webhook-cert\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.087312 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lc98\" (UniqueName: \"kubernetes.io/projected/beb78a59-5cba-42ca-a7bb-5351e77835b8-kube-api-access-5lc98\") pod \"infra-operator-controller-manager-7dc9c6fbcb-wzs5p\" (UID: \"beb78a59-5cba-42ca-a7bb-5351e77835b8\") " pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.213380 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.355889 4702 generic.go:334] "Generic (PLEG): container finished" podID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerID="dd2c783c3307a83295f9d1c5de644276ebab694aa4cf90cd6ac37e3b809a641a" exitCode=0 Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.356022 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkbfq" event={"ID":"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2","Type":"ContainerDied","Data":"dd2c783c3307a83295f9d1c5de644276ebab694aa4cf90cd6ac37e3b809a641a"} Oct 07 12:37:52 crc kubenswrapper[4702]: I1007 12:37:52.665515 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p"] Oct 07 12:37:52 crc kubenswrapper[4702]: W1007 12:37:52.672683 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeb78a59_5cba_42ca_a7bb_5351e77835b8.slice/crio-590866e2281076d5c356b84c271f12c41a16283352d44ab269a0a0ef74ea7e3d WatchSource:0}: Error finding container 590866e2281076d5c356b84c271f12c41a16283352d44ab269a0a0ef74ea7e3d: Status 404 returned error can't find the container with id 590866e2281076d5c356b84c271f12c41a16283352d44ab269a0a0ef74ea7e3d Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.342324 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krkwq"] Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.344211 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.355133 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krkwq"] Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.362459 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" event={"ID":"beb78a59-5cba-42ca-a7bb-5351e77835b8","Type":"ContainerStarted","Data":"590866e2281076d5c356b84c271f12c41a16283352d44ab269a0a0ef74ea7e3d"} Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.395398 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f95513-76af-43cd-ab53-5be28017d39d-utilities\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.395463 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k8xw\" (UniqueName: \"kubernetes.io/projected/49f95513-76af-43cd-ab53-5be28017d39d-kube-api-access-4k8xw\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.395511 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f95513-76af-43cd-ab53-5be28017d39d-catalog-content\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.497112 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f95513-76af-43cd-ab53-5be28017d39d-utilities\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.497193 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k8xw\" (UniqueName: \"kubernetes.io/projected/49f95513-76af-43cd-ab53-5be28017d39d-kube-api-access-4k8xw\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.497257 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f95513-76af-43cd-ab53-5be28017d39d-catalog-content\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.498144 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f95513-76af-43cd-ab53-5be28017d39d-catalog-content\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.498159 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f95513-76af-43cd-ab53-5be28017d39d-utilities\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.544838 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k8xw\" (UniqueName: \"kubernetes.io/projected/49f95513-76af-43cd-ab53-5be28017d39d-kube-api-access-4k8xw\") pod \"redhat-operators-krkwq\" (UID: \"49f95513-76af-43cd-ab53-5be28017d39d\") " pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.714494 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:37:53 crc kubenswrapper[4702]: I1007 12:37:53.959975 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krkwq"] Oct 07 12:37:53 crc kubenswrapper[4702]: W1007 12:37:53.979795 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49f95513_76af_43cd_ab53_5be28017d39d.slice/crio-7e100a61e8e2954e9d8fdc20e45130aa96961751cf47deaba971faa0a5b657bc WatchSource:0}: Error finding container 7e100a61e8e2954e9d8fdc20e45130aa96961751cf47deaba971faa0a5b657bc: Status 404 returned error can't find the container with id 7e100a61e8e2954e9d8fdc20e45130aa96961751cf47deaba971faa0a5b657bc Oct 07 12:37:54 crc kubenswrapper[4702]: I1007 12:37:54.377819 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkbfq" event={"ID":"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2","Type":"ContainerStarted","Data":"2b13adc21782ebaa938cbc785caec3419fc2664aaa8f0dcc589c7bd3e39effbe"} Oct 07 12:37:54 crc kubenswrapper[4702]: I1007 12:37:54.381777 4702 generic.go:334] "Generic (PLEG): container finished" podID="49f95513-76af-43cd-ab53-5be28017d39d" containerID="84d50b245d14c311acb0ca818614f8ff7f41868111bf6efa536348f41565088e" exitCode=0 Oct 07 12:37:54 crc kubenswrapper[4702]: I1007 12:37:54.381829 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krkwq" event={"ID":"49f95513-76af-43cd-ab53-5be28017d39d","Type":"ContainerDied","Data":"84d50b245d14c311acb0ca818614f8ff7f41868111bf6efa536348f41565088e"} Oct 07 12:37:54 crc kubenswrapper[4702]: I1007 12:37:54.381857 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krkwq" event={"ID":"49f95513-76af-43cd-ab53-5be28017d39d","Type":"ContainerStarted","Data":"7e100a61e8e2954e9d8fdc20e45130aa96961751cf47deaba971faa0a5b657bc"} Oct 07 12:37:54 crc kubenswrapper[4702]: I1007 12:37:54.401536 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kkbfq" podStartSLOduration=3.594806009 podStartE2EDuration="6.40151294s" podCreationTimestamp="2025-10-07 12:37:48 +0000 UTC" firstStartedPulling="2025-10-07 12:37:50.340946465 +0000 UTC m=+826.516682581" lastFinishedPulling="2025-10-07 12:37:53.147653386 +0000 UTC m=+829.323389512" observedRunningTime="2025-10-07 12:37:54.399001688 +0000 UTC m=+830.574737824" watchObservedRunningTime="2025-10-07 12:37:54.40151294 +0000 UTC m=+830.577249056" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.405529 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" event={"ID":"beb78a59-5cba-42ca-a7bb-5351e77835b8","Type":"ContainerStarted","Data":"bf72fcef2ee533613d81ffe62ab41b07bfa68d1470a784b2cf58912a55e8ec62"} Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.406336 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" event={"ID":"beb78a59-5cba-42ca-a7bb-5351e77835b8","Type":"ContainerStarted","Data":"59a7a61a63ea905bd7e73399ea0f933f8ebfe8b0652488d4ab22d0d026b4ed45"} Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.406362 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.444753 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" podStartSLOduration=2.680652796 podStartE2EDuration="6.444721133s" podCreationTimestamp="2025-10-07 12:37:51 +0000 UTC" firstStartedPulling="2025-10-07 12:37:52.679160322 +0000 UTC m=+828.854896438" lastFinishedPulling="2025-10-07 12:37:56.443228659 +0000 UTC m=+832.618964775" observedRunningTime="2025-10-07 12:37:57.441714177 +0000 UTC m=+833.617450313" watchObservedRunningTime="2025-10-07 12:37:57.444721133 +0000 UTC m=+833.620457249" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.542584 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.544121 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.547242 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.548172 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.548174 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.549664 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-glk6q" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.550257 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.550810 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.572100 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.573450 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.576622 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.578194 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.579984 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.615225 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.630126 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662121 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-operator-scripts\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662192 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/94961522-a5fc-43a6-a62b-2bba41a15be9-secrets\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662245 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-config-data-default\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662295 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/94961522-a5fc-43a6-a62b-2bba41a15be9-config-data-generated\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662316 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-kolla-config\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662436 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5msk\" (UniqueName: \"kubernetes.io/projected/94961522-a5fc-43a6-a62b-2bba41a15be9-kube-api-access-f5msk\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.662474 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.763444 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-operator-scripts\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.763988 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/94961522-a5fc-43a6-a62b-2bba41a15be9-secrets\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764016 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764040 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-config-data-default\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764112 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk5bx\" (UniqueName: \"kubernetes.io/projected/f1ba8472-9f12-45f8-9dc5-562c2b936b18-kube-api-access-tk5bx\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764136 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-config-data-default\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764153 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-operator-scripts\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764170 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1b04b98-3506-4aff-afd5-f9a2730576f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764208 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764228 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1ba8472-9f12-45f8-9dc5-562c2b936b18-config-data-generated\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764249 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/94961522-a5fc-43a6-a62b-2bba41a15be9-config-data-generated\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764267 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-kolla-config\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764288 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764306 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbp4k\" (UniqueName: \"kubernetes.io/projected/f1b04b98-3506-4aff-afd5-f9a2730576f6-kube-api-access-xbp4k\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764327 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764348 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5msk\" (UniqueName: \"kubernetes.io/projected/94961522-a5fc-43a6-a62b-2bba41a15be9-kube-api-access-f5msk\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764366 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764392 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-kolla-config\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764412 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764430 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1ba8472-9f12-45f8-9dc5-562c2b936b18-secrets\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.764444 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1b04b98-3506-4aff-afd5-f9a2730576f6-secrets\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.765165 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/94961522-a5fc-43a6-a62b-2bba41a15be9-config-data-generated\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.765360 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-operator-scripts\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.765352 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-config-data-default\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.765795 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94961522-a5fc-43a6-a62b-2bba41a15be9-kolla-config\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.765870 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.797194 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/94961522-a5fc-43a6-a62b-2bba41a15be9-secrets\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.797439 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5msk\" (UniqueName: \"kubernetes.io/projected/94961522-a5fc-43a6-a62b-2bba41a15be9-kube-api-access-f5msk\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.823176 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-1\" (UID: \"94961522-a5fc-43a6-a62b-2bba41a15be9\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866372 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866438 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1ba8472-9f12-45f8-9dc5-562c2b936b18-secrets\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866461 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1b04b98-3506-4aff-afd5-f9a2730576f6-secrets\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866498 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866527 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-config-data-default\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866552 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk5bx\" (UniqueName: \"kubernetes.io/projected/f1ba8472-9f12-45f8-9dc5-562c2b936b18-kube-api-access-tk5bx\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866577 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-operator-scripts\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866602 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1b04b98-3506-4aff-afd5-f9a2730576f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866647 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866672 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1ba8472-9f12-45f8-9dc5-562c2b936b18-config-data-generated\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866703 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866729 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbp4k\" (UniqueName: \"kubernetes.io/projected/f1b04b98-3506-4aff-afd5-f9a2730576f6-kube-api-access-xbp4k\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866757 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.866796 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-kolla-config\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.867793 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.867830 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-kolla-config\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.868240 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1b04b98-3506-4aff-afd5-f9a2730576f6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.868533 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-operator-scripts\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.868677 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.868807 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.869838 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.870658 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1ba8472-9f12-45f8-9dc5-562c2b936b18-config-data-generated\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.872647 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1ba8472-9f12-45f8-9dc5-562c2b936b18-config-data-default\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.876181 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1b04b98-3506-4aff-afd5-f9a2730576f6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.876515 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1ba8472-9f12-45f8-9dc5-562c2b936b18-secrets\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.876626 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1b04b98-3506-4aff-afd5-f9a2730576f6-secrets\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.895354 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.903246 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbp4k\" (UniqueName: \"kubernetes.io/projected/f1b04b98-3506-4aff-afd5-f9a2730576f6-kube-api-access-xbp4k\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.904664 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.905918 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk5bx\" (UniqueName: \"kubernetes.io/projected/f1ba8472-9f12-45f8-9dc5-562c2b936b18-kube-api-access-tk5bx\") pod \"openstack-galera-2\" (UID: \"f1ba8472-9f12-45f8-9dc5-562c2b936b18\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.907438 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f1b04b98-3506-4aff-afd5-f9a2730576f6\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:57 crc kubenswrapper[4702]: I1007 12:37:57.925579 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:37:58 crc kubenswrapper[4702]: I1007 12:37:58.171263 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:37:58 crc kubenswrapper[4702]: I1007 12:37:58.416524 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 07 12:37:58 crc kubenswrapper[4702]: W1007 12:37:58.437154 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94961522_a5fc_43a6_a62b_2bba41a15be9.slice/crio-a91a2d8a687e46514a5d479a3efcd538ab9807826744d5edd8e25809aff97727 WatchSource:0}: Error finding container a91a2d8a687e46514a5d479a3efcd538ab9807826744d5edd8e25809aff97727: Status 404 returned error can't find the container with id a91a2d8a687e46514a5d479a3efcd538ab9807826744d5edd8e25809aff97727 Oct 07 12:37:58 crc kubenswrapper[4702]: I1007 12:37:58.652153 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 07 12:37:58 crc kubenswrapper[4702]: W1007 12:37:58.655343 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1ba8472_9f12_45f8_9dc5_562c2b936b18.slice/crio-04437a0916add5b931c5972d3d5ae3de90b211dd858b8cfb596de62f2ea4eb42 WatchSource:0}: Error finding container 04437a0916add5b931c5972d3d5ae3de90b211dd858b8cfb596de62f2ea4eb42: Status 404 returned error can't find the container with id 04437a0916add5b931c5972d3d5ae3de90b211dd858b8cfb596de62f2ea4eb42 Oct 07 12:37:58 crc kubenswrapper[4702]: I1007 12:37:58.784224 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 07 12:37:58 crc kubenswrapper[4702]: W1007 12:37:58.791005 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1b04b98_3506_4aff_afd5_f9a2730576f6.slice/crio-3affebdaa91a275338d133e98806ff1908b533b86f1eecdd43bc656e17e7b2c0 WatchSource:0}: Error finding container 3affebdaa91a275338d133e98806ff1908b533b86f1eecdd43bc656e17e7b2c0: Status 404 returned error can't find the container with id 3affebdaa91a275338d133e98806ff1908b533b86f1eecdd43bc656e17e7b2c0 Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.150021 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.150531 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.198525 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.437495 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f1b04b98-3506-4aff-afd5-f9a2730576f6","Type":"ContainerStarted","Data":"3affebdaa91a275338d133e98806ff1908b533b86f1eecdd43bc656e17e7b2c0"} Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.438706 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"94961522-a5fc-43a6-a62b-2bba41a15be9","Type":"ContainerStarted","Data":"a91a2d8a687e46514a5d479a3efcd538ab9807826744d5edd8e25809aff97727"} Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.440366 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"f1ba8472-9f12-45f8-9dc5-562c2b936b18","Type":"ContainerStarted","Data":"04437a0916add5b931c5972d3d5ae3de90b211dd858b8cfb596de62f2ea4eb42"} Oct 07 12:37:59 crc kubenswrapper[4702]: I1007 12:37:59.484649 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:38:01 crc kubenswrapper[4702]: I1007 12:38:01.734514 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkbfq"] Oct 07 12:38:01 crc kubenswrapper[4702]: I1007 12:38:01.735123 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kkbfq" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="registry-server" containerID="cri-o://2b13adc21782ebaa938cbc785caec3419fc2664aaa8f0dcc589c7bd3e39effbe" gracePeriod=2 Oct 07 12:38:02 crc kubenswrapper[4702]: I1007 12:38:02.218663 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7dc9c6fbcb-wzs5p" Oct 07 12:38:02 crc kubenswrapper[4702]: I1007 12:38:02.480044 4702 generic.go:334] "Generic (PLEG): container finished" podID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerID="2b13adc21782ebaa938cbc785caec3419fc2664aaa8f0dcc589c7bd3e39effbe" exitCode=0 Oct 07 12:38:02 crc kubenswrapper[4702]: I1007 12:38:02.480108 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkbfq" event={"ID":"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2","Type":"ContainerDied","Data":"2b13adc21782ebaa938cbc785caec3419fc2664aaa8f0dcc589c7bd3e39effbe"} Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.359346 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.518106 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkbfq" event={"ID":"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2","Type":"ContainerDied","Data":"6efb27a7bdec3f8586530678b32308fd42369c30464434fd8a4bd6afcf9237d5"} Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.518176 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkbfq" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.518197 4702 scope.go:117] "RemoveContainer" containerID="2b13adc21782ebaa938cbc785caec3419fc2664aaa8f0dcc589c7bd3e39effbe" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.546653 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qdzg\" (UniqueName: \"kubernetes.io/projected/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-kube-api-access-7qdzg\") pod \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.546723 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-catalog-content\") pod \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.546844 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-utilities\") pod \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\" (UID: \"bfae9461-0b7a-4a21-acbb-cd74c5bc28e2\") " Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.548014 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-utilities" (OuterVolumeSpecName: "utilities") pod "bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" (UID: "bfae9461-0b7a-4a21-acbb-cd74c5bc28e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.573108 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-kube-api-access-7qdzg" (OuterVolumeSpecName: "kube-api-access-7qdzg") pod "bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" (UID: "bfae9461-0b7a-4a21-acbb-cd74c5bc28e2"). InnerVolumeSpecName "kube-api-access-7qdzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.613000 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" (UID: "bfae9461-0b7a-4a21-acbb-cd74c5bc28e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.648744 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.648792 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qdzg\" (UniqueName: \"kubernetes.io/projected/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-kube-api-access-7qdzg\") on node \"crc\" DevicePath \"\"" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.648805 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.844648 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkbfq"] Oct 07 12:38:07 crc kubenswrapper[4702]: I1007 12:38:07.858867 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kkbfq"] Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.544292 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-5gkxq"] Oct 07 12:38:08 crc kubenswrapper[4702]: E1007 12:38:08.544866 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="registry-server" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.544880 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="registry-server" Oct 07 12:38:08 crc kubenswrapper[4702]: E1007 12:38:08.544912 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="extract-utilities" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.544919 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="extract-utilities" Oct 07 12:38:08 crc kubenswrapper[4702]: E1007 12:38:08.544931 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="extract-content" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.544939 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="extract-content" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.545036 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" containerName="registry-server" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.545497 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.547787 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-mblzm" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.561212 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-5gkxq"] Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.651132 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfae9461-0b7a-4a21-acbb-cd74c5bc28e2" path="/var/lib/kubelet/pods/bfae9461-0b7a-4a21-acbb-cd74c5bc28e2/volumes" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.669866 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52xlq\" (UniqueName: \"kubernetes.io/projected/a21457b3-d5bd-4506-8bbc-50ff5579c2b4-kube-api-access-52xlq\") pod \"rabbitmq-cluster-operator-index-5gkxq\" (UID: \"a21457b3-d5bd-4506-8bbc-50ff5579c2b4\") " pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.772087 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52xlq\" (UniqueName: \"kubernetes.io/projected/a21457b3-d5bd-4506-8bbc-50ff5579c2b4-kube-api-access-52xlq\") pod \"rabbitmq-cluster-operator-index-5gkxq\" (UID: \"a21457b3-d5bd-4506-8bbc-50ff5579c2b4\") " pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.810338 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52xlq\" (UniqueName: \"kubernetes.io/projected/a21457b3-d5bd-4506-8bbc-50ff5579c2b4-kube-api-access-52xlq\") pod \"rabbitmq-cluster-operator-index-5gkxq\" (UID: \"a21457b3-d5bd-4506-8bbc-50ff5579c2b4\") " pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:08 crc kubenswrapper[4702]: I1007 12:38:08.864455 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.133850 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-5gkxq"] Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.745174 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-pphjk"] Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.748558 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.757407 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-pphjk"] Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.863113 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbcwd\" (UniqueName: \"kubernetes.io/projected/ec9bb282-6aca-4099-901d-b5bc2c780f43-kube-api-access-lbcwd\") pod \"rabbitmq-cluster-operator-index-pphjk\" (UID: \"ec9bb282-6aca-4099-901d-b5bc2c780f43\") " pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.964245 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbcwd\" (UniqueName: \"kubernetes.io/projected/ec9bb282-6aca-4099-901d-b5bc2c780f43-kube-api-access-lbcwd\") pod \"rabbitmq-cluster-operator-index-pphjk\" (UID: \"ec9bb282-6aca-4099-901d-b5bc2c780f43\") " pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:14 crc kubenswrapper[4702]: I1007 12:38:14.988075 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbcwd\" (UniqueName: \"kubernetes.io/projected/ec9bb282-6aca-4099-901d-b5bc2c780f43-kube-api-access-lbcwd\") pod \"rabbitmq-cluster-operator-index-pphjk\" (UID: \"ec9bb282-6aca-4099-901d-b5bc2c780f43\") " pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:15 crc kubenswrapper[4702]: I1007 12:38:15.077881 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:22 crc kubenswrapper[4702]: E1007 12:38:22.931585 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 07 12:38:22 crc kubenswrapper[4702]: E1007 12:38:22.933667 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4k8xw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-krkwq_openshift-marketplace(49f95513-76af-43cd-ab53-5be28017d39d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 12:38:22 crc kubenswrapper[4702]: E1007 12:38:22.935005 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-krkwq" podUID="49f95513-76af-43cd-ab53-5be28017d39d" Oct 07 12:38:22 crc kubenswrapper[4702]: I1007 12:38:22.996212 4702 scope.go:117] "RemoveContainer" containerID="dd2c783c3307a83295f9d1c5de644276ebab694aa4cf90cd6ac37e3b809a641a" Oct 07 12:38:23 crc kubenswrapper[4702]: I1007 12:38:23.142527 4702 scope.go:117] "RemoveContainer" containerID="e0843d540bdc986381de3039825519bf313c4ac3cac2a53f1e7c5432f2fd737a" Oct 07 12:38:23 crc kubenswrapper[4702]: I1007 12:38:23.340009 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-5gkxq"] Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.444163 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.444370 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tk5bx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-2_glance-kuttl-tests(f1ba8472-9f12-45f8-9dc5-562c2b936b18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.445608 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-2" podUID="f1ba8472-9f12-45f8-9dc5-562c2b936b18" Oct 07 12:38:23 crc kubenswrapper[4702]: I1007 12:38:23.627841 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-pphjk"] Oct 07 12:38:23 crc kubenswrapper[4702]: W1007 12:38:23.635858 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec9bb282_6aca_4099_901d_b5bc2c780f43.slice/crio-38e00b1a929e4ea23e4f5210cca48399e452ddded798378fcc7a969b758c27e0 WatchSource:0}: Error finding container 38e00b1a929e4ea23e4f5210cca48399e452ddded798378fcc7a969b758c27e0: Status 404 returned error can't find the container with id 38e00b1a929e4ea23e4f5210cca48399e452ddded798378fcc7a969b758c27e0 Oct 07 12:38:23 crc kubenswrapper[4702]: I1007 12:38:23.637777 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" event={"ID":"a21457b3-d5bd-4506-8bbc-50ff5579c2b4","Type":"ContainerStarted","Data":"a719b020d2f09e7a7b342cf675b87267124b1b068df8592170fc43566800da57"} Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.639282 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-krkwq" podUID="49f95513-76af-43cd-ab53-5be28017d39d" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.640371 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c\\\"\"" pod="glance-kuttl-tests/openstack-galera-2" podUID="f1ba8472-9f12-45f8-9dc5-562c2b936b18" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.728475 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.729076 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f5msk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-1_glance-kuttl-tests(94961522-a5fc-43a6-a62b-2bba41a15be9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.730273 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-1" podUID="94961522-a5fc-43a6-a62b-2bba41a15be9" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.858938 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.859244 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbp4k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_glance-kuttl-tests(f1b04b98-3506-4aff-afd5-f9a2730576f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 12:38:23 crc kubenswrapper[4702]: E1007 12:38:23.861154 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-0" podUID="f1b04b98-3506-4aff-afd5-f9a2730576f6" Oct 07 12:38:24 crc kubenswrapper[4702]: I1007 12:38:24.647392 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" event={"ID":"ec9bb282-6aca-4099-901d-b5bc2c780f43","Type":"ContainerStarted","Data":"38e00b1a929e4ea23e4f5210cca48399e452ddded798378fcc7a969b758c27e0"} Oct 07 12:38:24 crc kubenswrapper[4702]: E1007 12:38:24.649154 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c\\\"\"" pod="glance-kuttl-tests/openstack-galera-1" podUID="94961522-a5fc-43a6-a62b-2bba41a15be9" Oct 07 12:38:24 crc kubenswrapper[4702]: E1007 12:38:24.649221 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c\\\"\"" pod="glance-kuttl-tests/openstack-galera-0" podUID="f1b04b98-3506-4aff-afd5-f9a2730576f6" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.197102 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.197970 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.200555 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-8gr59" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.201025 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.211545 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.322732 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6cc5554b-8ac6-4893-bc94-3cfc4f815438-kolla-config\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.322781 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wwzv\" (UniqueName: \"kubernetes.io/projected/6cc5554b-8ac6-4893-bc94-3cfc4f815438-kube-api-access-5wwzv\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.322814 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cc5554b-8ac6-4893-bc94-3cfc4f815438-config-data\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.424745 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6cc5554b-8ac6-4893-bc94-3cfc4f815438-kolla-config\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.424803 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wwzv\" (UniqueName: \"kubernetes.io/projected/6cc5554b-8ac6-4893-bc94-3cfc4f815438-kube-api-access-5wwzv\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.424844 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cc5554b-8ac6-4893-bc94-3cfc4f815438-config-data\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.426176 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cc5554b-8ac6-4893-bc94-3cfc4f815438-config-data\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.426363 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6cc5554b-8ac6-4893-bc94-3cfc4f815438-kolla-config\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.455183 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wwzv\" (UniqueName: \"kubernetes.io/projected/6cc5554b-8ac6-4893-bc94-3cfc4f815438-kube-api-access-5wwzv\") pod \"memcached-0\" (UID: \"6cc5554b-8ac6-4893-bc94-3cfc4f815438\") " pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:25 crc kubenswrapper[4702]: I1007 12:38:25.520139 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:26 crc kubenswrapper[4702]: I1007 12:38:26.048155 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 07 12:38:26 crc kubenswrapper[4702]: W1007 12:38:26.056634 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cc5554b_8ac6_4893_bc94_3cfc4f815438.slice/crio-aa348c47bcf94b1b29c43d27ca172a5b7c9a31f51fae60c2ff72ba85880d49af WatchSource:0}: Error finding container aa348c47bcf94b1b29c43d27ca172a5b7c9a31f51fae60c2ff72ba85880d49af: Status 404 returned error can't find the container with id aa348c47bcf94b1b29c43d27ca172a5b7c9a31f51fae60c2ff72ba85880d49af Oct 07 12:38:26 crc kubenswrapper[4702]: I1007 12:38:26.709154 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"6cc5554b-8ac6-4893-bc94-3cfc4f815438","Type":"ContainerStarted","Data":"aa348c47bcf94b1b29c43d27ca172a5b7c9a31f51fae60c2ff72ba85880d49af"} Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.148882 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v5lg7"] Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.151079 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.164018 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5lg7"] Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.244631 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-catalog-content\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.245148 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dpn7\" (UniqueName: \"kubernetes.io/projected/9639c675-aec6-4104-9089-9b8a1f6e2785-kube-api-access-4dpn7\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.245189 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-utilities\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.346535 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-utilities\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.346724 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-catalog-content\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.346763 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dpn7\" (UniqueName: \"kubernetes.io/projected/9639c675-aec6-4104-9089-9b8a1f6e2785-kube-api-access-4dpn7\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.347315 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-utilities\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.347468 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-catalog-content\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.373966 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dpn7\" (UniqueName: \"kubernetes.io/projected/9639c675-aec6-4104-9089-9b8a1f6e2785-kube-api-access-4dpn7\") pod \"certified-operators-v5lg7\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.486651 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:38:32 crc kubenswrapper[4702]: I1007 12:38:32.967292 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5lg7"] Oct 07 12:38:32 crc kubenswrapper[4702]: W1007 12:38:32.983031 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9639c675_aec6_4104_9089_9b8a1f6e2785.slice/crio-1ca758f747c5e5e2a463b6db34d269048eff88c5e3a21b0fddd0319ff27b82e8 WatchSource:0}: Error finding container 1ca758f747c5e5e2a463b6db34d269048eff88c5e3a21b0fddd0319ff27b82e8: Status 404 returned error can't find the container with id 1ca758f747c5e5e2a463b6db34d269048eff88c5e3a21b0fddd0319ff27b82e8 Oct 07 12:38:33 crc kubenswrapper[4702]: I1007 12:38:33.756161 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5lg7" event={"ID":"9639c675-aec6-4104-9089-9b8a1f6e2785","Type":"ContainerStarted","Data":"1ca758f747c5e5e2a463b6db34d269048eff88c5e3a21b0fddd0319ff27b82e8"} Oct 07 12:38:34 crc kubenswrapper[4702]: I1007 12:38:34.764609 4702 generic.go:334] "Generic (PLEG): container finished" podID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerID="3938e7734c0823ff31979c8b40a9ff5871f5ef21919379d7853f93e15cd84fce" exitCode=0 Oct 07 12:38:34 crc kubenswrapper[4702]: I1007 12:38:34.764688 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5lg7" event={"ID":"9639c675-aec6-4104-9089-9b8a1f6e2785","Type":"ContainerDied","Data":"3938e7734c0823ff31979c8b40a9ff5871f5ef21919379d7853f93e15cd84fce"} Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.839360 4702 generic.go:334] "Generic (PLEG): container finished" podID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerID="f3d81cd7c7609d252c806a0be74e26e538716089625c4093ba5d48b76823aa57" exitCode=0 Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.839449 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5lg7" event={"ID":"9639c675-aec6-4104-9089-9b8a1f6e2785","Type":"ContainerDied","Data":"f3d81cd7c7609d252c806a0be74e26e538716089625c4093ba5d48b76823aa57"} Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.843113 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"94961522-a5fc-43a6-a62b-2bba41a15be9","Type":"ContainerStarted","Data":"4b8a735d572680ef234c1175c48e6e0247be96bc2630b5211374b4cf26a96f80"} Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.846046 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"f1ba8472-9f12-45f8-9dc5-562c2b936b18","Type":"ContainerStarted","Data":"d869bf8d63d6034c338657ba30cf4450e6b3d1e477d99fa2ae87ffc0e32548ac"} Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.847870 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"6cc5554b-8ac6-4893-bc94-3cfc4f815438","Type":"ContainerStarted","Data":"5596705946aaf33258d04af71cb8faa4e466e0d543fc2db952e78d026871eed7"} Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.848480 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:41 crc kubenswrapper[4702]: I1007 12:38:41.937551 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=2.903260351 podStartE2EDuration="16.937528813s" podCreationTimestamp="2025-10-07 12:38:25 +0000 UTC" firstStartedPulling="2025-10-07 12:38:26.060606002 +0000 UTC m=+862.236342118" lastFinishedPulling="2025-10-07 12:38:40.094874454 +0000 UTC m=+876.270610580" observedRunningTime="2025-10-07 12:38:41.932689873 +0000 UTC m=+878.108425989" watchObservedRunningTime="2025-10-07 12:38:41.937528813 +0000 UTC m=+878.113264919" Oct 07 12:38:51 crc kubenswrapper[4702]: I1007 12:38:50.521975 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Oct 07 12:38:51 crc kubenswrapper[4702]: E1007 12:38:51.601153 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest" Oct 07 12:38:51 crc kubenswrapper[4702]: E1007 12:38:51.601479 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-52xlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-index-5gkxq_openstack-operators(a21457b3-d5bd-4506-8bbc-50ff5579c2b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 12:38:51 crc kubenswrapper[4702]: E1007 12:38:51.602758 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" podUID="a21457b3-d5bd-4506-8bbc-50ff5579c2b4" Oct 07 12:38:51 crc kubenswrapper[4702]: I1007 12:38:51.919770 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krkwq" event={"ID":"49f95513-76af-43cd-ab53-5be28017d39d","Type":"ContainerStarted","Data":"e82c7832ec39036dfa1ca949668d340670e01741ac5c75a2b4e7f8a7464ef9c6"} Oct 07 12:38:51 crc kubenswrapper[4702]: I1007 12:38:51.921512 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f1b04b98-3506-4aff-afd5-f9a2730576f6","Type":"ContainerStarted","Data":"c26da7cc40e02e2ccea3a308aad1aff47b8c95d53883da751e1643de6a2b56bd"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.317406 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.349154 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52xlq\" (UniqueName: \"kubernetes.io/projected/a21457b3-d5bd-4506-8bbc-50ff5579c2b4-kube-api-access-52xlq\") pod \"a21457b3-d5bd-4506-8bbc-50ff5579c2b4\" (UID: \"a21457b3-d5bd-4506-8bbc-50ff5579c2b4\") " Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.356601 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a21457b3-d5bd-4506-8bbc-50ff5579c2b4-kube-api-access-52xlq" (OuterVolumeSpecName: "kube-api-access-52xlq") pod "a21457b3-d5bd-4506-8bbc-50ff5579c2b4" (UID: "a21457b3-d5bd-4506-8bbc-50ff5579c2b4"). InnerVolumeSpecName "kube-api-access-52xlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.451436 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52xlq\" (UniqueName: \"kubernetes.io/projected/a21457b3-d5bd-4506-8bbc-50ff5579c2b4-kube-api-access-52xlq\") on node \"crc\" DevicePath \"\"" Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.935326 4702 generic.go:334] "Generic (PLEG): container finished" podID="49f95513-76af-43cd-ab53-5be28017d39d" containerID="e82c7832ec39036dfa1ca949668d340670e01741ac5c75a2b4e7f8a7464ef9c6" exitCode=0 Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.935446 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krkwq" event={"ID":"49f95513-76af-43cd-ab53-5be28017d39d","Type":"ContainerDied","Data":"e82c7832ec39036dfa1ca949668d340670e01741ac5c75a2b4e7f8a7464ef9c6"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.940551 4702 generic.go:334] "Generic (PLEG): container finished" podID="94961522-a5fc-43a6-a62b-2bba41a15be9" containerID="4b8a735d572680ef234c1175c48e6e0247be96bc2630b5211374b4cf26a96f80" exitCode=0 Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.940735 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"94961522-a5fc-43a6-a62b-2bba41a15be9","Type":"ContainerDied","Data":"4b8a735d572680ef234c1175c48e6e0247be96bc2630b5211374b4cf26a96f80"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.946514 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" event={"ID":"ec9bb282-6aca-4099-901d-b5bc2c780f43","Type":"ContainerStarted","Data":"80d9f24c6ad069641d4118ed6e230f8977fa702b45b71ee570329e058d153383"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.949976 4702 generic.go:334] "Generic (PLEG): container finished" podID="f1ba8472-9f12-45f8-9dc5-562c2b936b18" containerID="d869bf8d63d6034c338657ba30cf4450e6b3d1e477d99fa2ae87ffc0e32548ac" exitCode=0 Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.950041 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"f1ba8472-9f12-45f8-9dc5-562c2b936b18","Type":"ContainerDied","Data":"d869bf8d63d6034c338657ba30cf4450e6b3d1e477d99fa2ae87ffc0e32548ac"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.952796 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5lg7" event={"ID":"9639c675-aec6-4104-9089-9b8a1f6e2785","Type":"ContainerStarted","Data":"1d112ce61fc1443aa16f81706cb85a25c484db098ef68e2fab2a54af4810ae09"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.955773 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" event={"ID":"a21457b3-d5bd-4506-8bbc-50ff5579c2b4","Type":"ContainerDied","Data":"a719b020d2f09e7a7b342cf675b87267124b1b068df8592170fc43566800da57"} Oct 07 12:38:52 crc kubenswrapper[4702]: I1007 12:38:52.956106 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-5gkxq" Oct 07 12:38:53 crc kubenswrapper[4702]: I1007 12:38:53.083188 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v5lg7" podStartSLOduration=4.442255671 podStartE2EDuration="21.083158415s" podCreationTimestamp="2025-10-07 12:38:32 +0000 UTC" firstStartedPulling="2025-10-07 12:38:35.772790443 +0000 UTC m=+871.948526549" lastFinishedPulling="2025-10-07 12:38:52.413693137 +0000 UTC m=+888.589429293" observedRunningTime="2025-10-07 12:38:53.077694917 +0000 UTC m=+889.253431053" watchObservedRunningTime="2025-10-07 12:38:53.083158415 +0000 UTC m=+889.258894521" Oct 07 12:38:53 crc kubenswrapper[4702]: I1007 12:38:53.083658 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" podStartSLOduration=10.310994554 podStartE2EDuration="39.083652379s" podCreationTimestamp="2025-10-07 12:38:14 +0000 UTC" firstStartedPulling="2025-10-07 12:38:23.641403582 +0000 UTC m=+859.817139728" lastFinishedPulling="2025-10-07 12:38:52.414061437 +0000 UTC m=+888.589797553" observedRunningTime="2025-10-07 12:38:53.058762389 +0000 UTC m=+889.234498515" watchObservedRunningTime="2025-10-07 12:38:53.083652379 +0000 UTC m=+889.259388495" Oct 07 12:38:53 crc kubenswrapper[4702]: I1007 12:38:53.106145 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-5gkxq"] Oct 07 12:38:53 crc kubenswrapper[4702]: I1007 12:38:53.113160 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-5gkxq"] Oct 07 12:38:53 crc kubenswrapper[4702]: I1007 12:38:53.966814 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"94961522-a5fc-43a6-a62b-2bba41a15be9","Type":"ContainerStarted","Data":"80ab207ed6eddcf40d7d43aa1ee0b5d4a5a2358175063f7e666a6616a7fe0cea"} Oct 07 12:38:54 crc kubenswrapper[4702]: I1007 12:38:54.650714 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a21457b3-d5bd-4506-8bbc-50ff5579c2b4" path="/var/lib/kubelet/pods/a21457b3-d5bd-4506-8bbc-50ff5579c2b4/volumes" Oct 07 12:38:54 crc kubenswrapper[4702]: I1007 12:38:54.983146 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"f1ba8472-9f12-45f8-9dc5-562c2b936b18","Type":"ContainerStarted","Data":"6da6dbce037ed3ada3ea8ee02159c2ec671b5fa0f2bc7d62c93b857eb91b0395"} Oct 07 12:38:55 crc kubenswrapper[4702]: I1007 12:38:55.078627 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:55 crc kubenswrapper[4702]: I1007 12:38:55.078714 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:55 crc kubenswrapper[4702]: I1007 12:38:55.119634 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:38:56 crc kubenswrapper[4702]: I1007 12:38:56.022820 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=18.585978072 podStartE2EDuration="1m0.0227942s" podCreationTimestamp="2025-10-07 12:37:56 +0000 UTC" firstStartedPulling="2025-10-07 12:37:58.658967802 +0000 UTC m=+834.834703918" lastFinishedPulling="2025-10-07 12:38:40.09578393 +0000 UTC m=+876.271520046" observedRunningTime="2025-10-07 12:38:56.017726104 +0000 UTC m=+892.193462220" watchObservedRunningTime="2025-10-07 12:38:56.0227942 +0000 UTC m=+892.198530316" Oct 07 12:38:56 crc kubenswrapper[4702]: I1007 12:38:56.996640 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krkwq" event={"ID":"49f95513-76af-43cd-ab53-5be28017d39d","Type":"ContainerStarted","Data":"4f85671c24c306bd32617a20731c193ce03c054352c550675e58f0e54b8a1721"} Oct 07 12:38:57 crc kubenswrapper[4702]: I1007 12:38:57.016664 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krkwq" podStartSLOduration=2.360749232 podStartE2EDuration="1m4.016640564s" podCreationTimestamp="2025-10-07 12:37:53 +0000 UTC" firstStartedPulling="2025-10-07 12:37:54.383401136 +0000 UTC m=+830.559137252" lastFinishedPulling="2025-10-07 12:38:56.039292468 +0000 UTC m=+892.215028584" observedRunningTime="2025-10-07 12:38:57.014572094 +0000 UTC m=+893.190308210" watchObservedRunningTime="2025-10-07 12:38:57.016640564 +0000 UTC m=+893.192376680" Oct 07 12:38:57 crc kubenswrapper[4702]: I1007 12:38:57.016988 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=19.363627001 podStartE2EDuration="1m1.016984184s" podCreationTimestamp="2025-10-07 12:37:56 +0000 UTC" firstStartedPulling="2025-10-07 12:37:58.442483459 +0000 UTC m=+834.618219575" lastFinishedPulling="2025-10-07 12:38:40.095840602 +0000 UTC m=+876.271576758" observedRunningTime="2025-10-07 12:38:56.045164468 +0000 UTC m=+892.220900584" watchObservedRunningTime="2025-10-07 12:38:57.016984184 +0000 UTC m=+893.192720300" Oct 07 12:38:57 crc kubenswrapper[4702]: I1007 12:38:57.906186 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:38:57 crc kubenswrapper[4702]: I1007 12:38:57.906259 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:38:57 crc kubenswrapper[4702]: I1007 12:38:57.926230 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:38:57 crc kubenswrapper[4702]: I1007 12:38:57.927043 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:39:02 crc kubenswrapper[4702]: I1007 12:39:02.487271 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:39:02 crc kubenswrapper[4702]: I1007 12:39:02.487754 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:39:02 crc kubenswrapper[4702]: I1007 12:39:02.546415 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:39:03 crc kubenswrapper[4702]: I1007 12:39:03.086277 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:39:03 crc kubenswrapper[4702]: I1007 12:39:03.715838 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:39:03 crc kubenswrapper[4702]: I1007 12:39:03.715944 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:39:03 crc kubenswrapper[4702]: I1007 12:39:03.777857 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:39:04 crc kubenswrapper[4702]: I1007 12:39:04.119316 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krkwq" Oct 07 12:39:05 crc kubenswrapper[4702]: I1007 12:39:05.117566 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-pphjk" Oct 07 12:39:06 crc kubenswrapper[4702]: I1007 12:39:06.735969 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5lg7"] Oct 07 12:39:06 crc kubenswrapper[4702]: I1007 12:39:06.736730 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v5lg7" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="registry-server" containerID="cri-o://1d112ce61fc1443aa16f81706cb85a25c484db098ef68e2fab2a54af4810ae09" gracePeriod=2 Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.074020 4702 generic.go:334] "Generic (PLEG): container finished" podID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerID="1d112ce61fc1443aa16f81706cb85a25c484db098ef68e2fab2a54af4810ae09" exitCode=0 Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.074083 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5lg7" event={"ID":"9639c675-aec6-4104-9089-9b8a1f6e2785","Type":"ContainerDied","Data":"1d112ce61fc1443aa16f81706cb85a25c484db098ef68e2fab2a54af4810ae09"} Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.795021 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krkwq"] Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.884719 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.993309 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dpn7\" (UniqueName: \"kubernetes.io/projected/9639c675-aec6-4104-9089-9b8a1f6e2785-kube-api-access-4dpn7\") pod \"9639c675-aec6-4104-9089-9b8a1f6e2785\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.993419 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-utilities\") pod \"9639c675-aec6-4104-9089-9b8a1f6e2785\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.993562 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-catalog-content\") pod \"9639c675-aec6-4104-9089-9b8a1f6e2785\" (UID: \"9639c675-aec6-4104-9089-9b8a1f6e2785\") " Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.994597 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-utilities" (OuterVolumeSpecName: "utilities") pod "9639c675-aec6-4104-9089-9b8a1f6e2785" (UID: "9639c675-aec6-4104-9089-9b8a1f6e2785"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:39:07 crc kubenswrapper[4702]: I1007 12:39:07.995603 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.003701 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9639c675-aec6-4104-9089-9b8a1f6e2785-kube-api-access-4dpn7" (OuterVolumeSpecName: "kube-api-access-4dpn7") pod "9639c675-aec6-4104-9089-9b8a1f6e2785" (UID: "9639c675-aec6-4104-9089-9b8a1f6e2785"). InnerVolumeSpecName "kube-api-access-4dpn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.043822 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9639c675-aec6-4104-9089-9b8a1f6e2785" (UID: "9639c675-aec6-4104-9089-9b8a1f6e2785"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.085004 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5lg7" event={"ID":"9639c675-aec6-4104-9089-9b8a1f6e2785","Type":"ContainerDied","Data":"1ca758f747c5e5e2a463b6db34d269048eff88c5e3a21b0fddd0319ff27b82e8"} Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.085073 4702 scope.go:117] "RemoveContainer" containerID="1d112ce61fc1443aa16f81706cb85a25c484db098ef68e2fab2a54af4810ae09" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.085111 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5lg7" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.096451 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dpn7\" (UniqueName: \"kubernetes.io/projected/9639c675-aec6-4104-9089-9b8a1f6e2785-kube-api-access-4dpn7\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.096593 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9639c675-aec6-4104-9089-9b8a1f6e2785-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.124871 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5lg7"] Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.132659 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v5lg7"] Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.133486 4702 scope.go:117] "RemoveContainer" containerID="f3d81cd7c7609d252c806a0be74e26e538716089625c4093ba5d48b76823aa57" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.165782 4702 scope.go:117] "RemoveContainer" containerID="3938e7734c0823ff31979c8b40a9ff5871f5ef21919379d7853f93e15cd84fce" Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.538347 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-82sgm"] Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.538752 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-82sgm" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="registry-server" containerID="cri-o://d67e9e4b2485b0d9807c469ef1e7b0a7018e4be99e90350410c7af0e24b4af31" gracePeriod=2 Oct 07 12:39:08 crc kubenswrapper[4702]: I1007 12:39:08.648433 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" path="/var/lib/kubelet/pods/9639c675-aec6-4104-9089-9b8a1f6e2785/volumes" Oct 07 12:39:09 crc kubenswrapper[4702]: I1007 12:39:09.094779 4702 generic.go:334] "Generic (PLEG): container finished" podID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerID="d67e9e4b2485b0d9807c469ef1e7b0a7018e4be99e90350410c7af0e24b4af31" exitCode=0 Oct 07 12:39:09 crc kubenswrapper[4702]: I1007 12:39:09.094870 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerDied","Data":"d67e9e4b2485b0d9807c469ef1e7b0a7018e4be99e90350410c7af0e24b4af31"} Oct 07 12:39:09 crc kubenswrapper[4702]: I1007 12:39:09.263530 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:39:09 crc kubenswrapper[4702]: I1007 12:39:09.263617 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.388640 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.534476 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-catalog-content\") pod \"35e85fb0-1e14-4990-a158-3f693778ddbe\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.534876 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6jzc\" (UniqueName: \"kubernetes.io/projected/35e85fb0-1e14-4990-a158-3f693778ddbe-kube-api-access-g6jzc\") pod \"35e85fb0-1e14-4990-a158-3f693778ddbe\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.534999 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-utilities\") pod \"35e85fb0-1e14-4990-a158-3f693778ddbe\" (UID: \"35e85fb0-1e14-4990-a158-3f693778ddbe\") " Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.536235 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-utilities" (OuterVolumeSpecName: "utilities") pod "35e85fb0-1e14-4990-a158-3f693778ddbe" (UID: "35e85fb0-1e14-4990-a158-3f693778ddbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.548303 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e85fb0-1e14-4990-a158-3f693778ddbe-kube-api-access-g6jzc" (OuterVolumeSpecName: "kube-api-access-g6jzc") pod "35e85fb0-1e14-4990-a158-3f693778ddbe" (UID: "35e85fb0-1e14-4990-a158-3f693778ddbe"). InnerVolumeSpecName "kube-api-access-g6jzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.637160 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6jzc\" (UniqueName: \"kubernetes.io/projected/35e85fb0-1e14-4990-a158-3f693778ddbe-kube-api-access-g6jzc\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.637205 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.638939 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35e85fb0-1e14-4990-a158-3f693778ddbe" (UID: "35e85fb0-1e14-4990-a158-3f693778ddbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:39:10 crc kubenswrapper[4702]: I1007 12:39:10.738406 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e85fb0-1e14-4990-a158-3f693778ddbe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.115088 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82sgm" event={"ID":"35e85fb0-1e14-4990-a158-3f693778ddbe","Type":"ContainerDied","Data":"49c1ee46f9c3c6fd1dd80fe6ce318796bd6fc35be3562a29d85188259767800e"} Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.115170 4702 scope.go:117] "RemoveContainer" containerID="d67e9e4b2485b0d9807c469ef1e7b0a7018e4be99e90350410c7af0e24b4af31" Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.115375 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82sgm" Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.146729 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-82sgm"] Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.148278 4702 scope.go:117] "RemoveContainer" containerID="1a7c044a1765b68b759f88538d04c91631f1c80e7cedc24521c2bf257004e8bf" Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.151786 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-82sgm"] Oct 07 12:39:11 crc kubenswrapper[4702]: I1007 12:39:11.171020 4702 scope.go:117] "RemoveContainer" containerID="dcbdb1fb358e0f48df841e674187ed5490bb2147c13e4396ed6fb70618ff36bc" Oct 07 12:39:12 crc kubenswrapper[4702]: I1007 12:39:12.650400 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" path="/var/lib/kubelet/pods/35e85fb0-1e14-4990-a158-3f693778ddbe/volumes" Oct 07 12:39:15 crc kubenswrapper[4702]: I1007 12:39:15.147660 4702 generic.go:334] "Generic (PLEG): container finished" podID="f1b04b98-3506-4aff-afd5-f9a2730576f6" containerID="c26da7cc40e02e2ccea3a308aad1aff47b8c95d53883da751e1643de6a2b56bd" exitCode=0 Oct 07 12:39:15 crc kubenswrapper[4702]: I1007 12:39:15.147738 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f1b04b98-3506-4aff-afd5-f9a2730576f6","Type":"ContainerDied","Data":"c26da7cc40e02e2ccea3a308aad1aff47b8c95d53883da751e1643de6a2b56bd"} Oct 07 12:39:16 crc kubenswrapper[4702]: I1007 12:39:16.158674 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f1b04b98-3506-4aff-afd5-f9a2730576f6","Type":"ContainerStarted","Data":"03cdf16599070703c62a52f75dbbe078917a271f4c55cadc7b4d97a1b0d1acf3"} Oct 07 12:39:16 crc kubenswrapper[4702]: I1007 12:39:16.188465 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=38.081597048 podStartE2EDuration="1m20.188431929s" podCreationTimestamp="2025-10-07 12:37:56 +0000 UTC" firstStartedPulling="2025-10-07 12:37:58.793868645 +0000 UTC m=+834.969604761" lastFinishedPulling="2025-10-07 12:38:40.900703516 +0000 UTC m=+877.076439642" observedRunningTime="2025-10-07 12:39:16.181634982 +0000 UTC m=+912.357371138" watchObservedRunningTime="2025-10-07 12:39:16.188431929 +0000 UTC m=+912.364168065" Oct 07 12:39:18 crc kubenswrapper[4702]: I1007 12:39:18.171542 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:39:18 crc kubenswrapper[4702]: I1007 12:39:18.173200 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:39:20 crc kubenswrapper[4702]: I1007 12:39:20.091650 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:39:20 crc kubenswrapper[4702]: I1007 12:39:20.153559 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="f1ba8472-9f12-45f8-9dc5-562c2b936b18" containerName="galera" probeResult="failure" output=< Oct 07 12:39:20 crc kubenswrapper[4702]: wsrep_local_state_comment (Joined) differs from Synced Oct 07 12:39:20 crc kubenswrapper[4702]: > Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.121766 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh"] Oct 07 12:39:23 crc kubenswrapper[4702]: E1007 12:39:23.123164 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="extract-content" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.123297 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="extract-content" Oct 07 12:39:23 crc kubenswrapper[4702]: E1007 12:39:23.123384 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="extract-utilities" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.123449 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="extract-utilities" Oct 07 12:39:23 crc kubenswrapper[4702]: E1007 12:39:23.123513 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="registry-server" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.123570 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="registry-server" Oct 07 12:39:23 crc kubenswrapper[4702]: E1007 12:39:23.123627 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="registry-server" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.123690 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="registry-server" Oct 07 12:39:23 crc kubenswrapper[4702]: E1007 12:39:23.123753 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="extract-utilities" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.123806 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="extract-utilities" Oct 07 12:39:23 crc kubenswrapper[4702]: E1007 12:39:23.123933 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="extract-content" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.123996 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="extract-content" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.124156 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="9639c675-aec6-4104-9089-9b8a1f6e2785" containerName="registry-server" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.124226 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e85fb0-1e14-4990-a158-3f693778ddbe" containerName="registry-server" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.125338 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.131459 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrk4d" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.144091 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh"] Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.220051 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.220122 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqkb8\" (UniqueName: \"kubernetes.io/projected/1bf04750-2468-4c4d-8931-b03f97a6b36a-kube-api-access-pqkb8\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.220140 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.321411 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.321529 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqkb8\" (UniqueName: \"kubernetes.io/projected/1bf04750-2468-4c4d-8931-b03f97a6b36a-kube-api-access-pqkb8\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.321557 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.322196 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.322287 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.346681 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqkb8\" (UniqueName: \"kubernetes.io/projected/1bf04750-2468-4c4d-8931-b03f97a6b36a-kube-api-access-pqkb8\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.449673 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:23 crc kubenswrapper[4702]: I1007 12:39:23.672682 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh"] Oct 07 12:39:24 crc kubenswrapper[4702]: I1007 12:39:24.213709 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerStarted","Data":"107854d57e528abd23a5078a25dce1c6db02c945805ee28c0eeb97fad4401819"} Oct 07 12:39:25 crc kubenswrapper[4702]: I1007 12:39:25.222421 4702 generic.go:334] "Generic (PLEG): container finished" podID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerID="779f33f46c785ad4d9488c7c6baabe20eb9afae01b0c2deffeef9d330eea1121" exitCode=0 Oct 07 12:39:25 crc kubenswrapper[4702]: I1007 12:39:25.222534 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerDied","Data":"779f33f46c785ad4d9488c7c6baabe20eb9afae01b0c2deffeef9d330eea1121"} Oct 07 12:39:27 crc kubenswrapper[4702]: I1007 12:39:27.984709 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Oct 07 12:39:28 crc kubenswrapper[4702]: I1007 12:39:28.248726 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerStarted","Data":"049c5a3b5696f9bf619a6d3282d58ee0654e1b42528e75f346b1176369729f99"} Oct 07 12:39:29 crc kubenswrapper[4702]: I1007 12:39:29.259020 4702 generic.go:334] "Generic (PLEG): container finished" podID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerID="049c5a3b5696f9bf619a6d3282d58ee0654e1b42528e75f346b1176369729f99" exitCode=0 Oct 07 12:39:29 crc kubenswrapper[4702]: I1007 12:39:29.259109 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerDied","Data":"049c5a3b5696f9bf619a6d3282d58ee0654e1b42528e75f346b1176369729f99"} Oct 07 12:39:30 crc kubenswrapper[4702]: I1007 12:39:30.271785 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerStarted","Data":"012908bc06cf16546b97ff91e0f893c58d3e958548a34906b56de5c3dc8d8f1e"} Oct 07 12:39:31 crc kubenswrapper[4702]: I1007 12:39:31.282306 4702 generic.go:334] "Generic (PLEG): container finished" podID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerID="012908bc06cf16546b97ff91e0f893c58d3e958548a34906b56de5c3dc8d8f1e" exitCode=0 Oct 07 12:39:31 crc kubenswrapper[4702]: I1007 12:39:31.282391 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerDied","Data":"012908bc06cf16546b97ff91e0f893c58d3e958548a34906b56de5c3dc8d8f1e"} Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.582780 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.768056 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-bundle\") pod \"1bf04750-2468-4c4d-8931-b03f97a6b36a\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.768106 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-util\") pod \"1bf04750-2468-4c4d-8931-b03f97a6b36a\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.768151 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqkb8\" (UniqueName: \"kubernetes.io/projected/1bf04750-2468-4c4d-8931-b03f97a6b36a-kube-api-access-pqkb8\") pod \"1bf04750-2468-4c4d-8931-b03f97a6b36a\" (UID: \"1bf04750-2468-4c4d-8931-b03f97a6b36a\") " Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.769405 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-bundle" (OuterVolumeSpecName: "bundle") pod "1bf04750-2468-4c4d-8931-b03f97a6b36a" (UID: "1bf04750-2468-4c4d-8931-b03f97a6b36a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.775164 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf04750-2468-4c4d-8931-b03f97a6b36a-kube-api-access-pqkb8" (OuterVolumeSpecName: "kube-api-access-pqkb8") pod "1bf04750-2468-4c4d-8931-b03f97a6b36a" (UID: "1bf04750-2468-4c4d-8931-b03f97a6b36a"). InnerVolumeSpecName "kube-api-access-pqkb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.784692 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-util" (OuterVolumeSpecName: "util") pod "1bf04750-2468-4c4d-8931-b03f97a6b36a" (UID: "1bf04750-2468-4c4d-8931-b03f97a6b36a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.869524 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqkb8\" (UniqueName: \"kubernetes.io/projected/1bf04750-2468-4c4d-8931-b03f97a6b36a-kube-api-access-pqkb8\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.869581 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:32 crc kubenswrapper[4702]: I1007 12:39:32.869597 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1bf04750-2468-4c4d-8931-b03f97a6b36a-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:39:33 crc kubenswrapper[4702]: I1007 12:39:33.302886 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" event={"ID":"1bf04750-2468-4c4d-8931-b03f97a6b36a","Type":"ContainerDied","Data":"107854d57e528abd23a5078a25dce1c6db02c945805ee28c0eeb97fad4401819"} Oct 07 12:39:33 crc kubenswrapper[4702]: I1007 12:39:33.302985 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="107854d57e528abd23a5078a25dce1c6db02c945805ee28c0eeb97fad4401819" Oct 07 12:39:33 crc kubenswrapper[4702]: I1007 12:39:33.303097 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh" Oct 07 12:39:38 crc kubenswrapper[4702]: I1007 12:39:38.314056 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:39:38 crc kubenswrapper[4702]: I1007 12:39:38.360936 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.263099 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.263186 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.337777 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn"] Oct 07 12:39:39 crc kubenswrapper[4702]: E1007 12:39:39.338164 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="pull" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.338187 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="pull" Oct 07 12:39:39 crc kubenswrapper[4702]: E1007 12:39:39.338206 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="extract" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.338217 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="extract" Oct 07 12:39:39 crc kubenswrapper[4702]: E1007 12:39:39.338255 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="util" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.338267 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="util" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.338454 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf04750-2468-4c4d-8931-b03f97a6b36a" containerName="extract" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.339143 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.346561 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-dpj7w" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.355663 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn"] Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.475146 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq8qz\" (UniqueName: \"kubernetes.io/projected/feb903e7-e1ea-4519-9913-f68dfb45c9e1-kube-api-access-dq8qz\") pod \"rabbitmq-cluster-operator-779fc9694b-zr2gn\" (UID: \"feb903e7-e1ea-4519-9913-f68dfb45c9e1\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.576551 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq8qz\" (UniqueName: \"kubernetes.io/projected/feb903e7-e1ea-4519-9913-f68dfb45c9e1-kube-api-access-dq8qz\") pod \"rabbitmq-cluster-operator-779fc9694b-zr2gn\" (UID: \"feb903e7-e1ea-4519-9913-f68dfb45c9e1\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.604179 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq8qz\" (UniqueName: \"kubernetes.io/projected/feb903e7-e1ea-4519-9913-f68dfb45c9e1-kube-api-access-dq8qz\") pod \"rabbitmq-cluster-operator-779fc9694b-zr2gn\" (UID: \"feb903e7-e1ea-4519-9913-f68dfb45c9e1\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.660079 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" Oct 07 12:39:39 crc kubenswrapper[4702]: I1007 12:39:39.895813 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn"] Oct 07 12:39:39 crc kubenswrapper[4702]: W1007 12:39:39.907065 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podfeb903e7_e1ea_4519_9913_f68dfb45c9e1.slice/crio-79bf681bbedd009f6cf5b172dc6ee1d7656e30e7562ce7c7b52eeb10eafd1dc2 WatchSource:0}: Error finding container 79bf681bbedd009f6cf5b172dc6ee1d7656e30e7562ce7c7b52eeb10eafd1dc2: Status 404 returned error can't find the container with id 79bf681bbedd009f6cf5b172dc6ee1d7656e30e7562ce7c7b52eeb10eafd1dc2 Oct 07 12:39:40 crc kubenswrapper[4702]: I1007 12:39:40.349155 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" event={"ID":"feb903e7-e1ea-4519-9913-f68dfb45c9e1","Type":"ContainerStarted","Data":"79bf681bbedd009f6cf5b172dc6ee1d7656e30e7562ce7c7b52eeb10eafd1dc2"} Oct 07 12:39:41 crc kubenswrapper[4702]: I1007 12:39:41.970069 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:39:42 crc kubenswrapper[4702]: I1007 12:39:42.029606 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Oct 07 12:39:42 crc kubenswrapper[4702]: I1007 12:39:42.364731 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" event={"ID":"feb903e7-e1ea-4519-9913-f68dfb45c9e1","Type":"ContainerStarted","Data":"fe4632abb6a81bcc33bf7f22cdd1a6afd11acb77d7840b97abc7b3379b9dac3a"} Oct 07 12:39:42 crc kubenswrapper[4702]: I1007 12:39:42.381685 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zr2gn" podStartSLOduration=1.621342128 podStartE2EDuration="3.381661701s" podCreationTimestamp="2025-10-07 12:39:39 +0000 UTC" firstStartedPulling="2025-10-07 12:39:39.90983359 +0000 UTC m=+936.085569706" lastFinishedPulling="2025-10-07 12:39:41.670153163 +0000 UTC m=+937.845889279" observedRunningTime="2025-10-07 12:39:42.379498919 +0000 UTC m=+938.555235025" watchObservedRunningTime="2025-10-07 12:39:42.381661701 +0000 UTC m=+938.557397817" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.421629 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.423165 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.425506 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.425560 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.425601 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.425613 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.425883 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-7wrqm" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.444947 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546083 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546149 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c3cf606b-7113-4066-8808-ebde118e2247-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546203 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546238 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c3cf606b-7113-4066-8808-ebde118e2247-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546325 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c3cf606b-7113-4066-8808-ebde118e2247-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546497 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546619 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.546680 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xlhv\" (UniqueName: \"kubernetes.io/projected/c3cf606b-7113-4066-8808-ebde118e2247-kube-api-access-5xlhv\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647300 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647347 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xlhv\" (UniqueName: \"kubernetes.io/projected/c3cf606b-7113-4066-8808-ebde118e2247-kube-api-access-5xlhv\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647368 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647391 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c3cf606b-7113-4066-8808-ebde118e2247-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647419 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647440 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c3cf606b-7113-4066-8808-ebde118e2247-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647495 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c3cf606b-7113-4066-8808-ebde118e2247-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.647522 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.648330 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.648721 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c3cf606b-7113-4066-8808-ebde118e2247-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.648925 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.655674 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c3cf606b-7113-4066-8808-ebde118e2247-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.656616 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c3cf606b-7113-4066-8808-ebde118e2247-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.658390 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c3cf606b-7113-4066-8808-ebde118e2247-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.661947 4702 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.661998 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7d79e4035d815df8b4c62d6b42c18f708a63b744f3cee181c7587fdf879af2f1/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.678809 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xlhv\" (UniqueName: \"kubernetes.io/projected/c3cf606b-7113-4066-8808-ebde118e2247-kube-api-access-5xlhv\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:44 crc kubenswrapper[4702]: I1007 12:39:44.754278 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ccfd1c-1971-4d0b-bebe-297b60bfad97\") pod \"rabbitmq-server-0\" (UID: \"c3cf606b-7113-4066-8808-ebde118e2247\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.041001 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.488114 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.866464 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-r6rq2"] Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.867612 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.870456 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-zgww9" Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.878383 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-r6rq2"] Oct 07 12:39:45 crc kubenswrapper[4702]: I1007 12:39:45.966494 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v595f\" (UniqueName: \"kubernetes.io/projected/1a57bec4-d2af-4a75-b137-e3bce61d7aef-kube-api-access-v595f\") pod \"keystone-operator-index-r6rq2\" (UID: \"1a57bec4-d2af-4a75-b137-e3bce61d7aef\") " pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:46 crc kubenswrapper[4702]: I1007 12:39:46.068442 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v595f\" (UniqueName: \"kubernetes.io/projected/1a57bec4-d2af-4a75-b137-e3bce61d7aef-kube-api-access-v595f\") pod \"keystone-operator-index-r6rq2\" (UID: \"1a57bec4-d2af-4a75-b137-e3bce61d7aef\") " pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:46 crc kubenswrapper[4702]: I1007 12:39:46.099210 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v595f\" (UniqueName: \"kubernetes.io/projected/1a57bec4-d2af-4a75-b137-e3bce61d7aef-kube-api-access-v595f\") pod \"keystone-operator-index-r6rq2\" (UID: \"1a57bec4-d2af-4a75-b137-e3bce61d7aef\") " pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:46 crc kubenswrapper[4702]: I1007 12:39:46.202376 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:46 crc kubenswrapper[4702]: I1007 12:39:46.400425 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"c3cf606b-7113-4066-8808-ebde118e2247","Type":"ContainerStarted","Data":"c3882798e399bb1f6f277a2f6700fb7cddd8031ab9f6d81d9319d465a9b192f7"} Oct 07 12:39:46 crc kubenswrapper[4702]: I1007 12:39:46.433645 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-r6rq2"] Oct 07 12:39:47 crc kubenswrapper[4702]: I1007 12:39:47.416683 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-r6rq2" event={"ID":"1a57bec4-d2af-4a75-b137-e3bce61d7aef","Type":"ContainerStarted","Data":"aec8a6d7ac311e656713d5d78a1ab6aaebd76d7f6efd8803eb67c936ef46f4eb"} Oct 07 12:39:48 crc kubenswrapper[4702]: I1007 12:39:48.426115 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-r6rq2" event={"ID":"1a57bec4-d2af-4a75-b137-e3bce61d7aef","Type":"ContainerStarted","Data":"70a3b962337c0d5e5539ee524efb2de189cce5160795118f0fdc14113b093681"} Oct 07 12:39:48 crc kubenswrapper[4702]: I1007 12:39:48.449257 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-r6rq2" podStartSLOduration=2.341686759 podStartE2EDuration="3.449222s" podCreationTimestamp="2025-10-07 12:39:45 +0000 UTC" firstStartedPulling="2025-10-07 12:39:46.456504302 +0000 UTC m=+942.632240418" lastFinishedPulling="2025-10-07 12:39:47.564039543 +0000 UTC m=+943.739775659" observedRunningTime="2025-10-07 12:39:48.446586634 +0000 UTC m=+944.622322770" watchObservedRunningTime="2025-10-07 12:39:48.449222 +0000 UTC m=+944.624958126" Oct 07 12:39:56 crc kubenswrapper[4702]: I1007 12:39:56.203141 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:56 crc kubenswrapper[4702]: I1007 12:39:56.203764 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:56 crc kubenswrapper[4702]: I1007 12:39:56.259783 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:56 crc kubenswrapper[4702]: I1007 12:39:56.550478 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-r6rq2" Oct 07 12:39:57 crc kubenswrapper[4702]: I1007 12:39:57.514868 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"c3cf606b-7113-4066-8808-ebde118e2247","Type":"ContainerStarted","Data":"0bc645006126394e14b88107e1b8587bfce3db8ce00660fb40937edb570c1425"} Oct 07 12:40:04 crc kubenswrapper[4702]: I1007 12:40:04.923433 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4"] Oct 07 12:40:04 crc kubenswrapper[4702]: I1007 12:40:04.925661 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:04 crc kubenswrapper[4702]: I1007 12:40:04.928386 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrk4d" Oct 07 12:40:04 crc kubenswrapper[4702]: I1007 12:40:04.949501 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4"] Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.060324 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-util\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.060980 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-bundle\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.061039 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvlmq\" (UniqueName: \"kubernetes.io/projected/96e269e9-1b43-40d4-8590-ee9290396d92-kube-api-access-jvlmq\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.162738 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-util\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.162916 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-bundle\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.163022 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvlmq\" (UniqueName: \"kubernetes.io/projected/96e269e9-1b43-40d4-8590-ee9290396d92-kube-api-access-jvlmq\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.163632 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-bundle\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.164032 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-util\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.196070 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvlmq\" (UniqueName: \"kubernetes.io/projected/96e269e9-1b43-40d4-8590-ee9290396d92-kube-api-access-jvlmq\") pod \"a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.243789 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:05 crc kubenswrapper[4702]: I1007 12:40:05.676916 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4"] Oct 07 12:40:06 crc kubenswrapper[4702]: I1007 12:40:06.579601 4702 generic.go:334] "Generic (PLEG): container finished" podID="96e269e9-1b43-40d4-8590-ee9290396d92" containerID="88042b34f0c1bd7f720e4a296fbf94fbf618780cca2d8a13de2fb6b2092a294d" exitCode=0 Oct 07 12:40:06 crc kubenswrapper[4702]: I1007 12:40:06.579665 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" event={"ID":"96e269e9-1b43-40d4-8590-ee9290396d92","Type":"ContainerDied","Data":"88042b34f0c1bd7f720e4a296fbf94fbf618780cca2d8a13de2fb6b2092a294d"} Oct 07 12:40:06 crc kubenswrapper[4702]: I1007 12:40:06.579709 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" event={"ID":"96e269e9-1b43-40d4-8590-ee9290396d92","Type":"ContainerStarted","Data":"45220926cd38e7ef3ce020a58d507056294adf0840077287f3a9e0a6f3ae955b"} Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.262991 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.264233 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.264328 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.265098 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ca59d66613e781014a1bca3d21d490470064dd363a5feb6c9d2c05e7748d68e"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.265158 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://3ca59d66613e781014a1bca3d21d490470064dd363a5feb6c9d2c05e7748d68e" gracePeriod=600 Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.605512 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="3ca59d66613e781014a1bca3d21d490470064dd363a5feb6c9d2c05e7748d68e" exitCode=0 Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.605723 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"3ca59d66613e781014a1bca3d21d490470064dd363a5feb6c9d2c05e7748d68e"} Oct 07 12:40:09 crc kubenswrapper[4702]: I1007 12:40:09.606467 4702 scope.go:117] "RemoveContainer" containerID="ef24fdfca720c04e2e8d0834767736417773f1bf6dd8396fbe53fc5a646c234f" Oct 07 12:40:10 crc kubenswrapper[4702]: I1007 12:40:10.617807 4702 generic.go:334] "Generic (PLEG): container finished" podID="96e269e9-1b43-40d4-8590-ee9290396d92" containerID="160a199601debc3a06937443db55429f761917ca9944d4d4fa6f709b4e069909" exitCode=0 Oct 07 12:40:10 crc kubenswrapper[4702]: I1007 12:40:10.617887 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" event={"ID":"96e269e9-1b43-40d4-8590-ee9290396d92","Type":"ContainerDied","Data":"160a199601debc3a06937443db55429f761917ca9944d4d4fa6f709b4e069909"} Oct 07 12:40:10 crc kubenswrapper[4702]: I1007 12:40:10.623259 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"03f130f70c653ad000f84ff22f5df3f9d3bfff20eccc689a18ef25482e6731a7"} Oct 07 12:40:11 crc kubenswrapper[4702]: I1007 12:40:11.633879 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" event={"ID":"96e269e9-1b43-40d4-8590-ee9290396d92","Type":"ContainerStarted","Data":"ce34a34bd6e83e6672c33d71e4f89ba4a25b4555ef6a00705da46db72c05e040"} Oct 07 12:40:11 crc kubenswrapper[4702]: I1007 12:40:11.670396 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" podStartSLOduration=4.868818152 podStartE2EDuration="7.670357593s" podCreationTimestamp="2025-10-07 12:40:04 +0000 UTC" firstStartedPulling="2025-10-07 12:40:06.582263076 +0000 UTC m=+962.757999192" lastFinishedPulling="2025-10-07 12:40:09.383802517 +0000 UTC m=+965.559538633" observedRunningTime="2025-10-07 12:40:11.661286439 +0000 UTC m=+967.837022575" watchObservedRunningTime="2025-10-07 12:40:11.670357593 +0000 UTC m=+967.846093739" Oct 07 12:40:12 crc kubenswrapper[4702]: I1007 12:40:12.648191 4702 generic.go:334] "Generic (PLEG): container finished" podID="96e269e9-1b43-40d4-8590-ee9290396d92" containerID="ce34a34bd6e83e6672c33d71e4f89ba4a25b4555ef6a00705da46db72c05e040" exitCode=0 Oct 07 12:40:12 crc kubenswrapper[4702]: I1007 12:40:12.657302 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" event={"ID":"96e269e9-1b43-40d4-8590-ee9290396d92","Type":"ContainerDied","Data":"ce34a34bd6e83e6672c33d71e4f89ba4a25b4555ef6a00705da46db72c05e040"} Oct 07 12:40:13 crc kubenswrapper[4702]: I1007 12:40:13.968605 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:13.997968 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-util\") pod \"96e269e9-1b43-40d4-8590-ee9290396d92\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:13.998028 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvlmq\" (UniqueName: \"kubernetes.io/projected/96e269e9-1b43-40d4-8590-ee9290396d92-kube-api-access-jvlmq\") pod \"96e269e9-1b43-40d4-8590-ee9290396d92\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:13.998075 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-bundle\") pod \"96e269e9-1b43-40d4-8590-ee9290396d92\" (UID: \"96e269e9-1b43-40d4-8590-ee9290396d92\") " Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.000082 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-bundle" (OuterVolumeSpecName: "bundle") pod "96e269e9-1b43-40d4-8590-ee9290396d92" (UID: "96e269e9-1b43-40d4-8590-ee9290396d92"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.021980 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-util" (OuterVolumeSpecName: "util") pod "96e269e9-1b43-40d4-8590-ee9290396d92" (UID: "96e269e9-1b43-40d4-8590-ee9290396d92"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.027347 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e269e9-1b43-40d4-8590-ee9290396d92-kube-api-access-jvlmq" (OuterVolumeSpecName: "kube-api-access-jvlmq") pod "96e269e9-1b43-40d4-8590-ee9290396d92" (UID: "96e269e9-1b43-40d4-8590-ee9290396d92"). InnerVolumeSpecName "kube-api-access-jvlmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.099265 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.099654 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvlmq\" (UniqueName: \"kubernetes.io/projected/96e269e9-1b43-40d4-8590-ee9290396d92-kube-api-access-jvlmq\") on node \"crc\" DevicePath \"\"" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.099719 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e269e9-1b43-40d4-8590-ee9290396d92-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.667618 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" event={"ID":"96e269e9-1b43-40d4-8590-ee9290396d92","Type":"ContainerDied","Data":"45220926cd38e7ef3ce020a58d507056294adf0840077287f3a9e0a6f3ae955b"} Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.667688 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45220926cd38e7ef3ce020a58d507056294adf0840077287f3a9e0a6f3ae955b" Oct 07 12:40:14 crc kubenswrapper[4702]: I1007 12:40:14.667696 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.498888 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf"] Oct 07 12:40:23 crc kubenswrapper[4702]: E1007 12:40:23.500241 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="extract" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.500265 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="extract" Oct 07 12:40:23 crc kubenswrapper[4702]: E1007 12:40:23.500308 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="pull" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.500317 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="pull" Oct 07 12:40:23 crc kubenswrapper[4702]: E1007 12:40:23.500330 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="util" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.500340 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="util" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.500495 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e269e9-1b43-40d4-8590-ee9290396d92" containerName="extract" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.501517 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.504151 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.504410 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-clb4l" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.511969 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf"] Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.637945 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t45hb\" (UniqueName: \"kubernetes.io/projected/444fbf26-f791-432e-8051-2bb3995dcd6f-kube-api-access-t45hb\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.638095 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/444fbf26-f791-432e-8051-2bb3995dcd6f-apiservice-cert\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.638138 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/444fbf26-f791-432e-8051-2bb3995dcd6f-webhook-cert\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.739580 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/444fbf26-f791-432e-8051-2bb3995dcd6f-apiservice-cert\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.739672 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/444fbf26-f791-432e-8051-2bb3995dcd6f-webhook-cert\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.739763 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t45hb\" (UniqueName: \"kubernetes.io/projected/444fbf26-f791-432e-8051-2bb3995dcd6f-kube-api-access-t45hb\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.747630 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/444fbf26-f791-432e-8051-2bb3995dcd6f-apiservice-cert\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.752770 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/444fbf26-f791-432e-8051-2bb3995dcd6f-webhook-cert\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.788244 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t45hb\" (UniqueName: \"kubernetes.io/projected/444fbf26-f791-432e-8051-2bb3995dcd6f-kube-api-access-t45hb\") pod \"keystone-operator-controller-manager-556f5598f7-x5kzf\" (UID: \"444fbf26-f791-432e-8051-2bb3995dcd6f\") " pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:23 crc kubenswrapper[4702]: I1007 12:40:23.827040 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:24 crc kubenswrapper[4702]: I1007 12:40:24.290373 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf"] Oct 07 12:40:24 crc kubenswrapper[4702]: W1007 12:40:24.300932 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod444fbf26_f791_432e_8051_2bb3995dcd6f.slice/crio-063a165009aa865fa513573b12ae282376ff052b0a73b9d27a7725bb2d80cc3c WatchSource:0}: Error finding container 063a165009aa865fa513573b12ae282376ff052b0a73b9d27a7725bb2d80cc3c: Status 404 returned error can't find the container with id 063a165009aa865fa513573b12ae282376ff052b0a73b9d27a7725bb2d80cc3c Oct 07 12:40:24 crc kubenswrapper[4702]: I1007 12:40:24.304754 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 12:40:24 crc kubenswrapper[4702]: I1007 12:40:24.753825 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" event={"ID":"444fbf26-f791-432e-8051-2bb3995dcd6f","Type":"ContainerStarted","Data":"063a165009aa865fa513573b12ae282376ff052b0a73b9d27a7725bb2d80cc3c"} Oct 07 12:40:29 crc kubenswrapper[4702]: I1007 12:40:29.792161 4702 generic.go:334] "Generic (PLEG): container finished" podID="c3cf606b-7113-4066-8808-ebde118e2247" containerID="0bc645006126394e14b88107e1b8587bfce3db8ce00660fb40937edb570c1425" exitCode=0 Oct 07 12:40:29 crc kubenswrapper[4702]: I1007 12:40:29.792282 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"c3cf606b-7113-4066-8808-ebde118e2247","Type":"ContainerDied","Data":"0bc645006126394e14b88107e1b8587bfce3db8ce00660fb40937edb570c1425"} Oct 07 12:40:30 crc kubenswrapper[4702]: I1007 12:40:30.802465 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"c3cf606b-7113-4066-8808-ebde118e2247","Type":"ContainerStarted","Data":"bbc06c91aae630637f28cb0c1c2370674dfb7f63714185096cbd1ea4f92218e5"} Oct 07 12:40:31 crc kubenswrapper[4702]: I1007 12:40:31.808556 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:40:31 crc kubenswrapper[4702]: I1007 12:40:31.834274 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=38.56040473 podStartE2EDuration="48.834241834s" podCreationTimestamp="2025-10-07 12:39:43 +0000 UTC" firstStartedPulling="2025-10-07 12:39:45.498284645 +0000 UTC m=+941.674020761" lastFinishedPulling="2025-10-07 12:39:55.772121749 +0000 UTC m=+951.947857865" observedRunningTime="2025-10-07 12:40:31.831404621 +0000 UTC m=+988.007140737" watchObservedRunningTime="2025-10-07 12:40:31.834241834 +0000 UTC m=+988.009977950" Oct 07 12:40:33 crc kubenswrapper[4702]: I1007 12:40:33.825051 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" event={"ID":"444fbf26-f791-432e-8051-2bb3995dcd6f","Type":"ContainerStarted","Data":"f1e219407e996c850dab7bed70e97d14f377d4e224b8794b38125b5207ef3f7c"} Oct 07 12:40:33 crc kubenswrapper[4702]: I1007 12:40:33.825557 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:33 crc kubenswrapper[4702]: I1007 12:40:33.825568 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" event={"ID":"444fbf26-f791-432e-8051-2bb3995dcd6f","Type":"ContainerStarted","Data":"70e08e5e60456706e2597fd573fe83ee4fba8090bb2f45370f8e9628f707deef"} Oct 07 12:40:33 crc kubenswrapper[4702]: I1007 12:40:33.865444 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" podStartSLOduration=2.017652858 podStartE2EDuration="10.865415469s" podCreationTimestamp="2025-10-07 12:40:23 +0000 UTC" firstStartedPulling="2025-10-07 12:40:24.304537323 +0000 UTC m=+980.480273439" lastFinishedPulling="2025-10-07 12:40:33.152299934 +0000 UTC m=+989.328036050" observedRunningTime="2025-10-07 12:40:33.861584197 +0000 UTC m=+990.037320333" watchObservedRunningTime="2025-10-07 12:40:33.865415469 +0000 UTC m=+990.041151585" Oct 07 12:40:43 crc kubenswrapper[4702]: I1007 12:40:43.832931 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-556f5598f7-x5kzf" Oct 07 12:40:45 crc kubenswrapper[4702]: I1007 12:40:45.046115 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.081826 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-p6frh"] Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.084181 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.093176 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-p6frh"] Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.144229 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g2cn\" (UniqueName: \"kubernetes.io/projected/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a-kube-api-access-2g2cn\") pod \"keystone-db-create-p6frh\" (UID: \"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a\") " pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.245803 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g2cn\" (UniqueName: \"kubernetes.io/projected/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a-kube-api-access-2g2cn\") pod \"keystone-db-create-p6frh\" (UID: \"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a\") " pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.278203 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g2cn\" (UniqueName: \"kubernetes.io/projected/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a-kube-api-access-2g2cn\") pod \"keystone-db-create-p6frh\" (UID: \"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a\") " pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.409268 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.882314 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-p6frh"] Oct 07 12:40:49 crc kubenswrapper[4702]: I1007 12:40:49.938778 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-p6frh" event={"ID":"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a","Type":"ContainerStarted","Data":"b7a1cb9214899e8d029bfe07b596def14a4776803523bae850ac45bb9c6e9163"} Oct 07 12:40:50 crc kubenswrapper[4702]: I1007 12:40:50.949059 4702 generic.go:334] "Generic (PLEG): container finished" podID="29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a" containerID="8024eddcc1b69fd6713bf0fe38024280c2862eb43138e4ca3019a4aed51f3c89" exitCode=0 Oct 07 12:40:50 crc kubenswrapper[4702]: I1007 12:40:50.949153 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-p6frh" event={"ID":"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a","Type":"ContainerDied","Data":"8024eddcc1b69fd6713bf0fe38024280c2862eb43138e4ca3019a4aed51f3c89"} Oct 07 12:40:51 crc kubenswrapper[4702]: I1007 12:40:51.866769 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-snplx"] Oct 07 12:40:51 crc kubenswrapper[4702]: I1007 12:40:51.867867 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:40:51 crc kubenswrapper[4702]: I1007 12:40:51.869965 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-4snh9" Oct 07 12:40:51 crc kubenswrapper[4702]: I1007 12:40:51.877190 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-snplx"] Oct 07 12:40:51 crc kubenswrapper[4702]: I1007 12:40:51.990084 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58mjz\" (UniqueName: \"kubernetes.io/projected/1032492a-6e8e-4795-ada8-1a3f84943f07-kube-api-access-58mjz\") pod \"horizon-operator-index-snplx\" (UID: \"1032492a-6e8e-4795-ada8-1a3f84943f07\") " pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.092769 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58mjz\" (UniqueName: \"kubernetes.io/projected/1032492a-6e8e-4795-ada8-1a3f84943f07-kube-api-access-58mjz\") pod \"horizon-operator-index-snplx\" (UID: \"1032492a-6e8e-4795-ada8-1a3f84943f07\") " pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.121337 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58mjz\" (UniqueName: \"kubernetes.io/projected/1032492a-6e8e-4795-ada8-1a3f84943f07-kube-api-access-58mjz\") pod \"horizon-operator-index-snplx\" (UID: \"1032492a-6e8e-4795-ada8-1a3f84943f07\") " pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.193120 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.263701 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.397003 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g2cn\" (UniqueName: \"kubernetes.io/projected/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a-kube-api-access-2g2cn\") pod \"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a\" (UID: \"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a\") " Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.403025 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a-kube-api-access-2g2cn" (OuterVolumeSpecName: "kube-api-access-2g2cn") pod "29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a" (UID: "29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a"). InnerVolumeSpecName "kube-api-access-2g2cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.438360 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-snplx"] Oct 07 12:40:52 crc kubenswrapper[4702]: W1007 12:40:52.444345 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1032492a_6e8e_4795_ada8_1a3f84943f07.slice/crio-a7a19d38b281528abcf6cd3ed839ba95530347e2be19b56829503551e307b186 WatchSource:0}: Error finding container a7a19d38b281528abcf6cd3ed839ba95530347e2be19b56829503551e307b186: Status 404 returned error can't find the container with id a7a19d38b281528abcf6cd3ed839ba95530347e2be19b56829503551e307b186 Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.498754 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g2cn\" (UniqueName: \"kubernetes.io/projected/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a-kube-api-access-2g2cn\") on node \"crc\" DevicePath \"\"" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.962757 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-snplx" event={"ID":"1032492a-6e8e-4795-ada8-1a3f84943f07","Type":"ContainerStarted","Data":"a7a19d38b281528abcf6cd3ed839ba95530347e2be19b56829503551e307b186"} Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.965089 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-p6frh" event={"ID":"29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a","Type":"ContainerDied","Data":"b7a1cb9214899e8d029bfe07b596def14a4776803523bae850ac45bb9c6e9163"} Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.965117 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7a1cb9214899e8d029bfe07b596def14a4776803523bae850ac45bb9c6e9163" Oct 07 12:40:52 crc kubenswrapper[4702]: I1007 12:40:52.965202 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-p6frh" Oct 07 12:40:53 crc kubenswrapper[4702]: I1007 12:40:53.973087 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-snplx" event={"ID":"1032492a-6e8e-4795-ada8-1a3f84943f07","Type":"ContainerStarted","Data":"22f4dd96cd5afdb2a1bfd2af68e282eb6ebef8a3c5c9745a8a28876552e10b77"} Oct 07 12:40:53 crc kubenswrapper[4702]: I1007 12:40:53.994936 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-snplx" podStartSLOduration=2.169214771 podStartE2EDuration="2.994900982s" podCreationTimestamp="2025-10-07 12:40:51 +0000 UTC" firstStartedPulling="2025-10-07 12:40:52.447069225 +0000 UTC m=+1008.622805341" lastFinishedPulling="2025-10-07 12:40:53.272755436 +0000 UTC m=+1009.448491552" observedRunningTime="2025-10-07 12:40:53.99138761 +0000 UTC m=+1010.167123736" watchObservedRunningTime="2025-10-07 12:40:53.994900982 +0000 UTC m=+1010.170637098" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.266225 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-j8lzr"] Oct 07 12:40:55 crc kubenswrapper[4702]: E1007 12:40:55.266505 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a" containerName="mariadb-database-create" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.266519 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a" containerName="mariadb-database-create" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.266645 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a" containerName="mariadb-database-create" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.267166 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.269965 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-vthsf" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.279953 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-j8lzr"] Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.340938 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqk74\" (UniqueName: \"kubernetes.io/projected/68a99f85-bca1-4976-a914-7f83cfb77ac4-kube-api-access-cqk74\") pod \"swift-operator-index-j8lzr\" (UID: \"68a99f85-bca1-4976-a914-7f83cfb77ac4\") " pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.442953 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqk74\" (UniqueName: \"kubernetes.io/projected/68a99f85-bca1-4976-a914-7f83cfb77ac4-kube-api-access-cqk74\") pod \"swift-operator-index-j8lzr\" (UID: \"68a99f85-bca1-4976-a914-7f83cfb77ac4\") " pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.464878 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqk74\" (UniqueName: \"kubernetes.io/projected/68a99f85-bca1-4976-a914-7f83cfb77ac4-kube-api-access-cqk74\") pod \"swift-operator-index-j8lzr\" (UID: \"68a99f85-bca1-4976-a914-7f83cfb77ac4\") " pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:40:55 crc kubenswrapper[4702]: I1007 12:40:55.604698 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:40:56 crc kubenswrapper[4702]: I1007 12:40:56.110132 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-j8lzr"] Oct 07 12:40:57 crc kubenswrapper[4702]: I1007 12:40:57.019118 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-j8lzr" event={"ID":"68a99f85-bca1-4976-a914-7f83cfb77ac4","Type":"ContainerStarted","Data":"507fb4cc465f4754e978f45bf6aec90adfb5bc4dd1468dc354ef1c8478110f85"} Oct 07 12:40:58 crc kubenswrapper[4702]: I1007 12:40:58.979181 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cccd-account-create-wb8v7"] Oct 07 12:40:58 crc kubenswrapper[4702]: I1007 12:40:58.981006 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:40:58 crc kubenswrapper[4702]: I1007 12:40:58.983850 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Oct 07 12:40:58 crc kubenswrapper[4702]: I1007 12:40:58.991251 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cccd-account-create-wb8v7"] Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.032943 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-j8lzr" event={"ID":"68a99f85-bca1-4976-a914-7f83cfb77ac4","Type":"ContainerStarted","Data":"3af30172ff09d7000199e5c3f7c83e0beb7de93a0e377f61fec915b21f4e5cab"} Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.058121 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-j8lzr" podStartSLOduration=1.884017675 podStartE2EDuration="4.058099226s" podCreationTimestamp="2025-10-07 12:40:55 +0000 UTC" firstStartedPulling="2025-10-07 12:40:56.122100244 +0000 UTC m=+1012.297836360" lastFinishedPulling="2025-10-07 12:40:58.296181785 +0000 UTC m=+1014.471917911" observedRunningTime="2025-10-07 12:40:59.051270628 +0000 UTC m=+1015.227006754" watchObservedRunningTime="2025-10-07 12:40:59.058099226 +0000 UTC m=+1015.233835342" Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.096803 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9n2b\" (UniqueName: \"kubernetes.io/projected/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1-kube-api-access-b9n2b\") pod \"keystone-cccd-account-create-wb8v7\" (UID: \"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1\") " pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.198953 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9n2b\" (UniqueName: \"kubernetes.io/projected/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1-kube-api-access-b9n2b\") pod \"keystone-cccd-account-create-wb8v7\" (UID: \"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1\") " pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.247997 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9n2b\" (UniqueName: \"kubernetes.io/projected/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1-kube-api-access-b9n2b\") pod \"keystone-cccd-account-create-wb8v7\" (UID: \"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1\") " pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.297796 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:40:59 crc kubenswrapper[4702]: I1007 12:40:59.586390 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cccd-account-create-wb8v7"] Oct 07 12:41:00 crc kubenswrapper[4702]: I1007 12:41:00.044338 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" event={"ID":"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1","Type":"ContainerStarted","Data":"eef78e54a2813c17fc204ba6096e8e15ebfa7991450f73a04ba75388f26c4652"} Oct 07 12:41:00 crc kubenswrapper[4702]: I1007 12:41:00.044410 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" event={"ID":"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1","Type":"ContainerStarted","Data":"ae2862129705f347bcb10b2692960cf6b0285c4a08f0007c8766c01601946fc4"} Oct 07 12:41:00 crc kubenswrapper[4702]: I1007 12:41:00.062101 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" podStartSLOduration=2.06207098 podStartE2EDuration="2.06207098s" podCreationTimestamp="2025-10-07 12:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:41:00.05965829 +0000 UTC m=+1016.235394406" watchObservedRunningTime="2025-10-07 12:41:00.06207098 +0000 UTC m=+1016.237807096" Oct 07 12:41:01 crc kubenswrapper[4702]: I1007 12:41:01.055944 4702 generic.go:334] "Generic (PLEG): container finished" podID="80c8e674-d14d-4594-b2c6-a6d3e92ab9f1" containerID="eef78e54a2813c17fc204ba6096e8e15ebfa7991450f73a04ba75388f26c4652" exitCode=0 Oct 07 12:41:01 crc kubenswrapper[4702]: I1007 12:41:01.056047 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" event={"ID":"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1","Type":"ContainerDied","Data":"eef78e54a2813c17fc204ba6096e8e15ebfa7991450f73a04ba75388f26c4652"} Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.193272 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.193867 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.237571 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.401648 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.447855 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9n2b\" (UniqueName: \"kubernetes.io/projected/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1-kube-api-access-b9n2b\") pod \"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1\" (UID: \"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1\") " Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.456738 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1-kube-api-access-b9n2b" (OuterVolumeSpecName: "kube-api-access-b9n2b") pod "80c8e674-d14d-4594-b2c6-a6d3e92ab9f1" (UID: "80c8e674-d14d-4594-b2c6-a6d3e92ab9f1"). InnerVolumeSpecName "kube-api-access-b9n2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:41:02 crc kubenswrapper[4702]: I1007 12:41:02.550336 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9n2b\" (UniqueName: \"kubernetes.io/projected/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1-kube-api-access-b9n2b\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:03 crc kubenswrapper[4702]: I1007 12:41:03.071761 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" event={"ID":"80c8e674-d14d-4594-b2c6-a6d3e92ab9f1","Type":"ContainerDied","Data":"ae2862129705f347bcb10b2692960cf6b0285c4a08f0007c8766c01601946fc4"} Oct 07 12:41:03 crc kubenswrapper[4702]: I1007 12:41:03.072345 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae2862129705f347bcb10b2692960cf6b0285c4a08f0007c8766c01601946fc4" Oct 07 12:41:03 crc kubenswrapper[4702]: I1007 12:41:03.071794 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cccd-account-create-wb8v7" Oct 07 12:41:03 crc kubenswrapper[4702]: I1007 12:41:03.114412 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-snplx" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.456464 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-8dlbs"] Oct 07 12:41:04 crc kubenswrapper[4702]: E1007 12:41:04.457264 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c8e674-d14d-4594-b2c6-a6d3e92ab9f1" containerName="mariadb-account-create" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.457282 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c8e674-d14d-4594-b2c6-a6d3e92ab9f1" containerName="mariadb-account-create" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.457479 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c8e674-d14d-4594-b2c6-a6d3e92ab9f1" containerName="mariadb-account-create" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.458223 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.460767 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.461098 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-bdjdn" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.463533 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.463620 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.468951 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-8dlbs"] Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.579338 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snn8f\" (UniqueName: \"kubernetes.io/projected/3e6332da-21f3-4136-94e9-c5302ba865ac-kube-api-access-snn8f\") pod \"keystone-db-sync-8dlbs\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.579509 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e6332da-21f3-4136-94e9-c5302ba865ac-config-data\") pod \"keystone-db-sync-8dlbs\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.680646 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snn8f\" (UniqueName: \"kubernetes.io/projected/3e6332da-21f3-4136-94e9-c5302ba865ac-kube-api-access-snn8f\") pod \"keystone-db-sync-8dlbs\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.680728 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e6332da-21f3-4136-94e9-c5302ba865ac-config-data\") pod \"keystone-db-sync-8dlbs\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.683423 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.700549 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snn8f\" (UniqueName: \"kubernetes.io/projected/3e6332da-21f3-4136-94e9-c5302ba865ac-kube-api-access-snn8f\") pod \"keystone-db-sync-8dlbs\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.702244 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e6332da-21f3-4136-94e9-c5302ba865ac-config-data\") pod \"keystone-db-sync-8dlbs\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.783556 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-bdjdn" Oct 07 12:41:04 crc kubenswrapper[4702]: I1007 12:41:04.791591 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:41:05 crc kubenswrapper[4702]: I1007 12:41:05.240634 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-8dlbs"] Oct 07 12:41:05 crc kubenswrapper[4702]: W1007 12:41:05.245253 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e6332da_21f3_4136_94e9_c5302ba865ac.slice/crio-4e14d5beaeb2dccf8e1c99fb397873c94989d1dcea598c79379a9993a6c129e3 WatchSource:0}: Error finding container 4e14d5beaeb2dccf8e1c99fb397873c94989d1dcea598c79379a9993a6c129e3: Status 404 returned error can't find the container with id 4e14d5beaeb2dccf8e1c99fb397873c94989d1dcea598c79379a9993a6c129e3 Oct 07 12:41:05 crc kubenswrapper[4702]: I1007 12:41:05.605518 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:41:05 crc kubenswrapper[4702]: I1007 12:41:05.605622 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:41:05 crc kubenswrapper[4702]: I1007 12:41:05.657389 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:41:06 crc kubenswrapper[4702]: I1007 12:41:06.091980 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" event={"ID":"3e6332da-21f3-4136-94e9-c5302ba865ac","Type":"ContainerStarted","Data":"4e14d5beaeb2dccf8e1c99fb397873c94989d1dcea598c79379a9993a6c129e3"} Oct 07 12:41:06 crc kubenswrapper[4702]: I1007 12:41:06.122367 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-j8lzr" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.131117 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp"] Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.138766 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.142255 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrk4d" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.142453 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp"] Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.203125 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfjw9\" (UniqueName: \"kubernetes.io/projected/e819cdf1-7dc2-436f-9d41-c9f4c2783957-kube-api-access-cfjw9\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.203182 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-bundle\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.203754 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-util\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.305314 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfjw9\" (UniqueName: \"kubernetes.io/projected/e819cdf1-7dc2-436f-9d41-c9f4c2783957-kube-api-access-cfjw9\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.305384 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-bundle\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.305486 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-util\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.306156 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-bundle\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.306240 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-util\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.326450 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfjw9\" (UniqueName: \"kubernetes.io/projected/e819cdf1-7dc2-436f-9d41-c9f4c2783957-kube-api-access-cfjw9\") pod \"ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:12 crc kubenswrapper[4702]: I1007 12:41:12.498564 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.110086 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf"] Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.112530 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.115646 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf"] Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.219026 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x28wr\" (UniqueName: \"kubernetes.io/projected/d73554d4-7d8e-4f1b-a49f-866642b8822c-kube-api-access-x28wr\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.219076 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-util\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.219153 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-bundle\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.320277 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x28wr\" (UniqueName: \"kubernetes.io/projected/d73554d4-7d8e-4f1b-a49f-866642b8822c-kube-api-access-x28wr\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.320321 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-util\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.320368 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-bundle\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.320890 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-bundle\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.322556 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-util\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.346515 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x28wr\" (UniqueName: \"kubernetes.io/projected/d73554d4-7d8e-4f1b-a49f-866642b8822c-kube-api-access-x28wr\") pod \"f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:13 crc kubenswrapper[4702]: I1007 12:41:13.440006 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:16 crc kubenswrapper[4702]: I1007 12:41:16.237129 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp"] Oct 07 12:41:16 crc kubenswrapper[4702]: I1007 12:41:16.380983 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf"] Oct 07 12:41:16 crc kubenswrapper[4702]: W1007 12:41:16.388829 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd73554d4_7d8e_4f1b_a49f_866642b8822c.slice/crio-60102ddd3ae53a0dbaf1aff154e396bc58e567f84f92406f32658d5820478e4d WatchSource:0}: Error finding container 60102ddd3ae53a0dbaf1aff154e396bc58e567f84f92406f32658d5820478e4d: Status 404 returned error can't find the container with id 60102ddd3ae53a0dbaf1aff154e396bc58e567f84f92406f32658d5820478e4d Oct 07 12:41:17 crc kubenswrapper[4702]: I1007 12:41:17.175317 4702 generic.go:334] "Generic (PLEG): container finished" podID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerID="28fb97c02b2c71c6d4e60efdfe9944334679160ff8d40a6d49ac30bb7974daa7" exitCode=0 Oct 07 12:41:17 crc kubenswrapper[4702]: I1007 12:41:17.175437 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" event={"ID":"e819cdf1-7dc2-436f-9d41-c9f4c2783957","Type":"ContainerDied","Data":"28fb97c02b2c71c6d4e60efdfe9944334679160ff8d40a6d49ac30bb7974daa7"} Oct 07 12:41:17 crc kubenswrapper[4702]: I1007 12:41:17.175923 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" event={"ID":"e819cdf1-7dc2-436f-9d41-c9f4c2783957","Type":"ContainerStarted","Data":"0546541f4b8c4dc7846eb5e6d5e08f7f228fe386539156bdddd8eba0f477edb6"} Oct 07 12:41:17 crc kubenswrapper[4702]: I1007 12:41:17.178231 4702 generic.go:334] "Generic (PLEG): container finished" podID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerID="55db9ecccdca1e82bac7767c1c266d9c8e639be4001458fb1e49a65e81bbbf85" exitCode=0 Oct 07 12:41:17 crc kubenswrapper[4702]: I1007 12:41:17.178280 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" event={"ID":"d73554d4-7d8e-4f1b-a49f-866642b8822c","Type":"ContainerDied","Data":"55db9ecccdca1e82bac7767c1c266d9c8e639be4001458fb1e49a65e81bbbf85"} Oct 07 12:41:17 crc kubenswrapper[4702]: I1007 12:41:17.178311 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" event={"ID":"d73554d4-7d8e-4f1b-a49f-866642b8822c","Type":"ContainerStarted","Data":"60102ddd3ae53a0dbaf1aff154e396bc58e567f84f92406f32658d5820478e4d"} Oct 07 12:41:21 crc kubenswrapper[4702]: I1007 12:41:21.225026 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" event={"ID":"3e6332da-21f3-4136-94e9-c5302ba865ac","Type":"ContainerStarted","Data":"629e04090a084a22efea3b46523dec96fd0e1c8b526acf9c768142b67ce59c5a"} Oct 07 12:41:21 crc kubenswrapper[4702]: I1007 12:41:21.257289 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" podStartSLOduration=2.49007857 podStartE2EDuration="17.257258781s" podCreationTimestamp="2025-10-07 12:41:04 +0000 UTC" firstStartedPulling="2025-10-07 12:41:05.24956569 +0000 UTC m=+1021.425301816" lastFinishedPulling="2025-10-07 12:41:20.016745891 +0000 UTC m=+1036.192482027" observedRunningTime="2025-10-07 12:41:21.252501303 +0000 UTC m=+1037.428237429" watchObservedRunningTime="2025-10-07 12:41:21.257258781 +0000 UTC m=+1037.432994897" Oct 07 12:41:28 crc kubenswrapper[4702]: I1007 12:41:28.284669 4702 generic.go:334] "Generic (PLEG): container finished" podID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerID="e0377485556a266f31d78425f21ce10b9a2dba44a664cfa52924dd9628e8eaaf" exitCode=0 Oct 07 12:41:28 crc kubenswrapper[4702]: I1007 12:41:28.284737 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" event={"ID":"e819cdf1-7dc2-436f-9d41-c9f4c2783957","Type":"ContainerDied","Data":"e0377485556a266f31d78425f21ce10b9a2dba44a664cfa52924dd9628e8eaaf"} Oct 07 12:41:29 crc kubenswrapper[4702]: I1007 12:41:29.297161 4702 generic.go:334] "Generic (PLEG): container finished" podID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerID="ccda7bfeaf632d519ad171004ab77aa5dfa31956c3aeb4238f2bfc006da2219b" exitCode=0 Oct 07 12:41:29 crc kubenswrapper[4702]: I1007 12:41:29.297273 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" event={"ID":"e819cdf1-7dc2-436f-9d41-c9f4c2783957","Type":"ContainerDied","Data":"ccda7bfeaf632d519ad171004ab77aa5dfa31956c3aeb4238f2bfc006da2219b"} Oct 07 12:41:29 crc kubenswrapper[4702]: I1007 12:41:29.300949 4702 generic.go:334] "Generic (PLEG): container finished" podID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerID="febc16bcf31115dc36f925acfc0733e2571f1661f22031a5e66c9f9890931816" exitCode=0 Oct 07 12:41:29 crc kubenswrapper[4702]: I1007 12:41:29.301055 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" event={"ID":"d73554d4-7d8e-4f1b-a49f-866642b8822c","Type":"ContainerDied","Data":"febc16bcf31115dc36f925acfc0733e2571f1661f22031a5e66c9f9890931816"} Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.313595 4702 generic.go:334] "Generic (PLEG): container finished" podID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerID="8bdf2dfc55c749990e51535dfa8f9d62eab4b69197238cca71e712d139e35924" exitCode=0 Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.313726 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" event={"ID":"d73554d4-7d8e-4f1b-a49f-866642b8822c","Type":"ContainerDied","Data":"8bdf2dfc55c749990e51535dfa8f9d62eab4b69197238cca71e712d139e35924"} Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.599827 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.793049 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-bundle\") pod \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.793153 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-util\") pod \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.793245 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfjw9\" (UniqueName: \"kubernetes.io/projected/e819cdf1-7dc2-436f-9d41-c9f4c2783957-kube-api-access-cfjw9\") pod \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\" (UID: \"e819cdf1-7dc2-436f-9d41-c9f4c2783957\") " Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.794359 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-bundle" (OuterVolumeSpecName: "bundle") pod "e819cdf1-7dc2-436f-9d41-c9f4c2783957" (UID: "e819cdf1-7dc2-436f-9d41-c9f4c2783957"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.795013 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.800684 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e819cdf1-7dc2-436f-9d41-c9f4c2783957-kube-api-access-cfjw9" (OuterVolumeSpecName: "kube-api-access-cfjw9") pod "e819cdf1-7dc2-436f-9d41-c9f4c2783957" (UID: "e819cdf1-7dc2-436f-9d41-c9f4c2783957"). InnerVolumeSpecName "kube-api-access-cfjw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.808516 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-util" (OuterVolumeSpecName: "util") pod "e819cdf1-7dc2-436f-9d41-c9f4c2783957" (UID: "e819cdf1-7dc2-436f-9d41-c9f4c2783957"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.897247 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e819cdf1-7dc2-436f-9d41-c9f4c2783957-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:30 crc kubenswrapper[4702]: I1007 12:41:30.897311 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfjw9\" (UniqueName: \"kubernetes.io/projected/e819cdf1-7dc2-436f-9d41-c9f4c2783957-kube-api-access-cfjw9\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.329649 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" event={"ID":"e819cdf1-7dc2-436f-9d41-c9f4c2783957","Type":"ContainerDied","Data":"0546541f4b8c4dc7846eb5e6d5e08f7f228fe386539156bdddd8eba0f477edb6"} Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.329750 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0546541f4b8c4dc7846eb5e6d5e08f7f228fe386539156bdddd8eba0f477edb6" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.329699 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.618283 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.807802 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-util\") pod \"d73554d4-7d8e-4f1b-a49f-866642b8822c\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.808058 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-bundle\") pod \"d73554d4-7d8e-4f1b-a49f-866642b8822c\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.808111 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x28wr\" (UniqueName: \"kubernetes.io/projected/d73554d4-7d8e-4f1b-a49f-866642b8822c-kube-api-access-x28wr\") pod \"d73554d4-7d8e-4f1b-a49f-866642b8822c\" (UID: \"d73554d4-7d8e-4f1b-a49f-866642b8822c\") " Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.809333 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-bundle" (OuterVolumeSpecName: "bundle") pod "d73554d4-7d8e-4f1b-a49f-866642b8822c" (UID: "d73554d4-7d8e-4f1b-a49f-866642b8822c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.824939 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73554d4-7d8e-4f1b-a49f-866642b8822c-kube-api-access-x28wr" (OuterVolumeSpecName: "kube-api-access-x28wr") pod "d73554d4-7d8e-4f1b-a49f-866642b8822c" (UID: "d73554d4-7d8e-4f1b-a49f-866642b8822c"). InnerVolumeSpecName "kube-api-access-x28wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.831099 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-util" (OuterVolumeSpecName: "util") pod "d73554d4-7d8e-4f1b-a49f-866642b8822c" (UID: "d73554d4-7d8e-4f1b-a49f-866642b8822c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.910327 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.910384 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x28wr\" (UniqueName: \"kubernetes.io/projected/d73554d4-7d8e-4f1b-a49f-866642b8822c-kube-api-access-x28wr\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:31 crc kubenswrapper[4702]: I1007 12:41:31.910398 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d73554d4-7d8e-4f1b-a49f-866642b8822c-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:41:32 crc kubenswrapper[4702]: I1007 12:41:32.341419 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" event={"ID":"d73554d4-7d8e-4f1b-a49f-866642b8822c","Type":"ContainerDied","Data":"60102ddd3ae53a0dbaf1aff154e396bc58e567f84f92406f32658d5820478e4d"} Oct 07 12:41:32 crc kubenswrapper[4702]: I1007 12:41:32.341472 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf" Oct 07 12:41:32 crc kubenswrapper[4702]: I1007 12:41:32.341486 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60102ddd3ae53a0dbaf1aff154e396bc58e567f84f92406f32658d5820478e4d" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.751446 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l"] Oct 07 12:41:42 crc kubenswrapper[4702]: E1007 12:41:42.752541 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="util" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752557 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="util" Oct 07 12:41:42 crc kubenswrapper[4702]: E1007 12:41:42.752597 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="util" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752610 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="util" Oct 07 12:41:42 crc kubenswrapper[4702]: E1007 12:41:42.752627 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="extract" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752641 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="extract" Oct 07 12:41:42 crc kubenswrapper[4702]: E1007 12:41:42.752655 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="pull" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752663 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="pull" Oct 07 12:41:42 crc kubenswrapper[4702]: E1007 12:41:42.752676 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="extract" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752684 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="extract" Oct 07 12:41:42 crc kubenswrapper[4702]: E1007 12:41:42.752695 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="pull" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752702 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="pull" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.752873 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="d73554d4-7d8e-4f1b-a49f-866642b8822c" containerName="extract" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.753008 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e819cdf1-7dc2-436f-9d41-c9f4c2783957" containerName="extract" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.754023 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.760114 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-kzcvm" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.760617 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.779087 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l"] Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.873520 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-apiservice-cert\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.873573 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-webhook-cert\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.873681 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxgk4\" (UniqueName: \"kubernetes.io/projected/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-kube-api-access-cxgk4\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.975245 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxgk4\" (UniqueName: \"kubernetes.io/projected/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-kube-api-access-cxgk4\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.975346 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-apiservice-cert\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.975373 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-webhook-cert\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.985851 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-webhook-cert\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.992769 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-apiservice-cert\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:42 crc kubenswrapper[4702]: I1007 12:41:42.998035 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxgk4\" (UniqueName: \"kubernetes.io/projected/1a4a43d0-4ae6-4ccf-927a-014abbb8be62-kube-api-access-cxgk4\") pod \"swift-operator-controller-manager-6c7959756-6fx5l\" (UID: \"1a4a43d0-4ae6-4ccf-927a-014abbb8be62\") " pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:43 crc kubenswrapper[4702]: I1007 12:41:43.074552 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:43 crc kubenswrapper[4702]: I1007 12:41:43.637527 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l"] Oct 07 12:41:44 crc kubenswrapper[4702]: I1007 12:41:44.460532 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" event={"ID":"1a4a43d0-4ae6-4ccf-927a-014abbb8be62","Type":"ContainerStarted","Data":"3ee93697500ad01af9d05631fa7b2118132558e33e5fe248407279e589bd7977"} Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.076864 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc"] Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.079121 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.082328 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-plts8" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.083830 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.096292 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc"] Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.131901 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bcc3148d-161a-46ea-910c-aec0e3b3c824-apiservice-cert\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.132095 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5g4v\" (UniqueName: \"kubernetes.io/projected/bcc3148d-161a-46ea-910c-aec0e3b3c824-kube-api-access-c5g4v\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.132166 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bcc3148d-161a-46ea-910c-aec0e3b3c824-webhook-cert\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.233468 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5g4v\" (UniqueName: \"kubernetes.io/projected/bcc3148d-161a-46ea-910c-aec0e3b3c824-kube-api-access-c5g4v\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.233522 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bcc3148d-161a-46ea-910c-aec0e3b3c824-webhook-cert\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.233595 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bcc3148d-161a-46ea-910c-aec0e3b3c824-apiservice-cert\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.243688 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bcc3148d-161a-46ea-910c-aec0e3b3c824-apiservice-cert\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.245465 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bcc3148d-161a-46ea-910c-aec0e3b3c824-webhook-cert\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.255406 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5g4v\" (UniqueName: \"kubernetes.io/projected/bcc3148d-161a-46ea-910c-aec0e3b3c824-kube-api-access-c5g4v\") pod \"horizon-operator-controller-manager-7f9cd5475c-4k5dc\" (UID: \"bcc3148d-161a-46ea-910c-aec0e3b3c824\") " pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.405869 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:41:53 crc kubenswrapper[4702]: I1007 12:41:53.659382 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc"] Oct 07 12:41:55 crc kubenswrapper[4702]: W1007 12:41:55.226874 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcc3148d_161a_46ea_910c_aec0e3b3c824.slice/crio-13c4bbff076edac2be5c2c01fccc72c141afe13b844b35e183193ddcdef13fb6 WatchSource:0}: Error finding container 13c4bbff076edac2be5c2c01fccc72c141afe13b844b35e183193ddcdef13fb6: Status 404 returned error can't find the container with id 13c4bbff076edac2be5c2c01fccc72c141afe13b844b35e183193ddcdef13fb6 Oct 07 12:41:55 crc kubenswrapper[4702]: I1007 12:41:55.552091 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" event={"ID":"bcc3148d-161a-46ea-910c-aec0e3b3c824","Type":"ContainerStarted","Data":"13c4bbff076edac2be5c2c01fccc72c141afe13b844b35e183193ddcdef13fb6"} Oct 07 12:41:58 crc kubenswrapper[4702]: I1007 12:41:58.588667 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" event={"ID":"1a4a43d0-4ae6-4ccf-927a-014abbb8be62","Type":"ContainerStarted","Data":"aa0769d076b2aa2908729d60f3c3710a3cafd1ef3ab0bb75897d717957b720ec"} Oct 07 12:41:59 crc kubenswrapper[4702]: I1007 12:41:59.601956 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" event={"ID":"1a4a43d0-4ae6-4ccf-927a-014abbb8be62","Type":"ContainerStarted","Data":"7bb69bc7d894a38d03279b019e1479a3ad249bb1f49c0a486d9c10a30009f5bd"} Oct 07 12:41:59 crc kubenswrapper[4702]: I1007 12:41:59.602582 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:41:59 crc kubenswrapper[4702]: I1007 12:41:59.627089 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" podStartSLOduration=3.938485392 podStartE2EDuration="17.627062452s" podCreationTimestamp="2025-10-07 12:41:42 +0000 UTC" firstStartedPulling="2025-10-07 12:41:43.653364968 +0000 UTC m=+1059.829101084" lastFinishedPulling="2025-10-07 12:41:57.341942028 +0000 UTC m=+1073.517678144" observedRunningTime="2025-10-07 12:41:59.622277444 +0000 UTC m=+1075.798013580" watchObservedRunningTime="2025-10-07 12:41:59.627062452 +0000 UTC m=+1075.802798578" Oct 07 12:42:01 crc kubenswrapper[4702]: I1007 12:42:01.624528 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" event={"ID":"bcc3148d-161a-46ea-910c-aec0e3b3c824","Type":"ContainerStarted","Data":"75d358f630fa7aa4037495dccc53d6e0a1b1260b3decf1eba392382d776aefb7"} Oct 07 12:42:03 crc kubenswrapper[4702]: I1007 12:42:03.081096 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6c7959756-6fx5l" Oct 07 12:42:03 crc kubenswrapper[4702]: I1007 12:42:03.648448 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" event={"ID":"bcc3148d-161a-46ea-910c-aec0e3b3c824","Type":"ContainerStarted","Data":"6f771bf156b5ab0c23287520646d206f69549743bf84a67999f4b6cb67b0136a"} Oct 07 12:42:03 crc kubenswrapper[4702]: I1007 12:42:03.648656 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.939731 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" podStartSLOduration=8.079285112 podStartE2EDuration="13.939707338s" podCreationTimestamp="2025-10-07 12:41:53 +0000 UTC" firstStartedPulling="2025-10-07 12:41:55.230008247 +0000 UTC m=+1071.405744383" lastFinishedPulling="2025-10-07 12:42:01.090430463 +0000 UTC m=+1077.266166609" observedRunningTime="2025-10-07 12:42:03.687171326 +0000 UTC m=+1079.862907492" watchObservedRunningTime="2025-10-07 12:42:06.939707338 +0000 UTC m=+1083.115443444" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.945180 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.950756 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.954865 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.955103 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-kvz66" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.955267 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.954867 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Oct 07 12:42:06 crc kubenswrapper[4702]: I1007 12:42:06.981124 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.075407 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7646503c-ea03-4ce3-9438-54b187e1aec3-cache\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.075460 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.075489 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjk78\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-kube-api-access-tjk78\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.075526 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7646503c-ea03-4ce3-9438-54b187e1aec3-lock\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.075840 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177106 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7646503c-ea03-4ce3-9438-54b187e1aec3-lock\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177201 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177261 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7646503c-ea03-4ce3-9438-54b187e1aec3-cache\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177297 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177330 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjk78\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-kube-api-access-tjk78\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: E1007 12:42:07.177479 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 07 12:42:07 crc kubenswrapper[4702]: E1007 12:42:07.177512 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 07 12:42:07 crc kubenswrapper[4702]: E1007 12:42:07.177588 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift podName:7646503c-ea03-4ce3-9438-54b187e1aec3 nodeName:}" failed. No retries permitted until 2025-10-07 12:42:07.677558991 +0000 UTC m=+1083.853295107 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift") pod "swift-storage-0" (UID: "7646503c-ea03-4ce3-9438-54b187e1aec3") : configmap "swift-ring-files" not found Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177780 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.177882 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7646503c-ea03-4ce3-9438-54b187e1aec3-lock\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.178144 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7646503c-ea03-4ce3-9438-54b187e1aec3-cache\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.219002 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.222114 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjk78\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-kube-api-access-tjk78\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.486463 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-gv9w7"] Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.490610 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.493718 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.493731 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.494736 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.505973 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-gv9w7"] Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.583869 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-swiftconf\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.583994 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54cvr\" (UniqueName: \"kubernetes.io/projected/5d1fc69f-1999-40a0-acf6-eba215843669-kube-api-access-54cvr\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.584036 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-scripts\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.584067 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d1fc69f-1999-40a0-acf6-eba215843669-etc-swift\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.584110 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-ring-data-devices\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.584335 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-dispersionconf\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.685785 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54cvr\" (UniqueName: \"kubernetes.io/projected/5d1fc69f-1999-40a0-acf6-eba215843669-kube-api-access-54cvr\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.685852 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-scripts\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.685886 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d1fc69f-1999-40a0-acf6-eba215843669-etc-swift\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.685968 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-ring-data-devices\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.686029 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-dispersionconf\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.686098 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.686150 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-swiftconf\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: E1007 12:42:07.686305 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 07 12:42:07 crc kubenswrapper[4702]: E1007 12:42:07.686343 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 07 12:42:07 crc kubenswrapper[4702]: E1007 12:42:07.686431 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift podName:7646503c-ea03-4ce3-9438-54b187e1aec3 nodeName:}" failed. No retries permitted until 2025-10-07 12:42:08.686392678 +0000 UTC m=+1084.862128954 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift") pod "swift-storage-0" (UID: "7646503c-ea03-4ce3-9438-54b187e1aec3") : configmap "swift-ring-files" not found Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.686784 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d1fc69f-1999-40a0-acf6-eba215843669-etc-swift\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.687052 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-scripts\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.687477 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-ring-data-devices\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.697806 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-swiftconf\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.697859 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-dispersionconf\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.720190 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54cvr\" (UniqueName: \"kubernetes.io/projected/5d1fc69f-1999-40a0-acf6-eba215843669-kube-api-access-54cvr\") pod \"swift-ring-rebalance-gv9w7\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:07 crc kubenswrapper[4702]: I1007 12:42:07.808350 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.055793 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-gv9w7"] Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.465671 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-r4dwv"] Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.466782 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.469676 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-m45kv" Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.476240 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-r4dwv"] Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.603059 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbv96\" (UniqueName: \"kubernetes.io/projected/bde85707-06b8-407a-9f37-51d6be9c666e-kube-api-access-gbv96\") pod \"glance-operator-index-r4dwv\" (UID: \"bde85707-06b8-407a-9f37-51d6be9c666e\") " pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.696305 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" event={"ID":"5d1fc69f-1999-40a0-acf6-eba215843669","Type":"ContainerStarted","Data":"b9e61485775b9102b82e5e558543ed154ac6d17b0b04eccb8dafabbb335c7810"} Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.705080 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbv96\" (UniqueName: \"kubernetes.io/projected/bde85707-06b8-407a-9f37-51d6be9c666e-kube-api-access-gbv96\") pod \"glance-operator-index-r4dwv\" (UID: \"bde85707-06b8-407a-9f37-51d6be9c666e\") " pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.705457 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:08 crc kubenswrapper[4702]: E1007 12:42:08.707023 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 07 12:42:08 crc kubenswrapper[4702]: E1007 12:42:08.707061 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 07 12:42:08 crc kubenswrapper[4702]: E1007 12:42:08.707132 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift podName:7646503c-ea03-4ce3-9438-54b187e1aec3 nodeName:}" failed. No retries permitted until 2025-10-07 12:42:10.707111128 +0000 UTC m=+1086.882847244 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift") pod "swift-storage-0" (UID: "7646503c-ea03-4ce3-9438-54b187e1aec3") : configmap "swift-ring-files" not found Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.745433 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbv96\" (UniqueName: \"kubernetes.io/projected/bde85707-06b8-407a-9f37-51d6be9c666e-kube-api-access-gbv96\") pod \"glance-operator-index-r4dwv\" (UID: \"bde85707-06b8-407a-9f37-51d6be9c666e\") " pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:08 crc kubenswrapper[4702]: I1007 12:42:08.796728 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:09 crc kubenswrapper[4702]: I1007 12:42:09.262639 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:42:09 crc kubenswrapper[4702]: I1007 12:42:09.263053 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:42:09 crc kubenswrapper[4702]: I1007 12:42:09.286747 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-r4dwv"] Oct 07 12:42:09 crc kubenswrapper[4702]: W1007 12:42:09.296365 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbde85707_06b8_407a_9f37_51d6be9c666e.slice/crio-99887b7e3e37aca29eda22873f4c86f6abacd5dcc082a3975345bf83b90c8c65 WatchSource:0}: Error finding container 99887b7e3e37aca29eda22873f4c86f6abacd5dcc082a3975345bf83b90c8c65: Status 404 returned error can't find the container with id 99887b7e3e37aca29eda22873f4c86f6abacd5dcc082a3975345bf83b90c8c65 Oct 07 12:42:09 crc kubenswrapper[4702]: I1007 12:42:09.710430 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-r4dwv" event={"ID":"bde85707-06b8-407a-9f37-51d6be9c666e","Type":"ContainerStarted","Data":"99887b7e3e37aca29eda22873f4c86f6abacd5dcc082a3975345bf83b90c8c65"} Oct 07 12:42:10 crc kubenswrapper[4702]: I1007 12:42:10.720150 4702 generic.go:334] "Generic (PLEG): container finished" podID="3e6332da-21f3-4136-94e9-c5302ba865ac" containerID="629e04090a084a22efea3b46523dec96fd0e1c8b526acf9c768142b67ce59c5a" exitCode=0 Oct 07 12:42:10 crc kubenswrapper[4702]: I1007 12:42:10.720239 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" event={"ID":"3e6332da-21f3-4136-94e9-c5302ba865ac","Type":"ContainerDied","Data":"629e04090a084a22efea3b46523dec96fd0e1c8b526acf9c768142b67ce59c5a"} Oct 07 12:42:10 crc kubenswrapper[4702]: I1007 12:42:10.744289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:10 crc kubenswrapper[4702]: E1007 12:42:10.744505 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 07 12:42:10 crc kubenswrapper[4702]: E1007 12:42:10.744526 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 07 12:42:10 crc kubenswrapper[4702]: E1007 12:42:10.744679 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift podName:7646503c-ea03-4ce3-9438-54b187e1aec3 nodeName:}" failed. No retries permitted until 2025-10-07 12:42:14.744655787 +0000 UTC m=+1090.920391893 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift") pod "swift-storage-0" (UID: "7646503c-ea03-4ce3-9438-54b187e1aec3") : configmap "swift-ring-files" not found Oct 07 12:42:12 crc kubenswrapper[4702]: I1007 12:42:12.868691 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-r4dwv"] Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.414256 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7f9cd5475c-4k5dc" Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.480930 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-lg4mj"] Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.483339 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.488607 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-lg4mj"] Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.593043 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrsm5\" (UniqueName: \"kubernetes.io/projected/63310df2-8f1e-4ddc-b41f-f338b754812e-kube-api-access-hrsm5\") pod \"glance-operator-index-lg4mj\" (UID: \"63310df2-8f1e-4ddc-b41f-f338b754812e\") " pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.695037 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrsm5\" (UniqueName: \"kubernetes.io/projected/63310df2-8f1e-4ddc-b41f-f338b754812e-kube-api-access-hrsm5\") pod \"glance-operator-index-lg4mj\" (UID: \"63310df2-8f1e-4ddc-b41f-f338b754812e\") " pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.722494 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrsm5\" (UniqueName: \"kubernetes.io/projected/63310df2-8f1e-4ddc-b41f-f338b754812e-kube-api-access-hrsm5\") pod \"glance-operator-index-lg4mj\" (UID: \"63310df2-8f1e-4ddc-b41f-f338b754812e\") " pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:13 crc kubenswrapper[4702]: I1007 12:42:13.811398 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:14 crc kubenswrapper[4702]: I1007 12:42:14.813734 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:14 crc kubenswrapper[4702]: E1007 12:42:14.813991 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 07 12:42:14 crc kubenswrapper[4702]: E1007 12:42:14.814221 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 07 12:42:14 crc kubenswrapper[4702]: E1007 12:42:14.814301 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift podName:7646503c-ea03-4ce3-9438-54b187e1aec3 nodeName:}" failed. No retries permitted until 2025-10-07 12:42:22.814275086 +0000 UTC m=+1098.990011202 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift") pod "swift-storage-0" (UID: "7646503c-ea03-4ce3-9438-54b187e1aec3") : configmap "swift-ring-files" not found Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.501366 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.524638 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e6332da-21f3-4136-94e9-c5302ba865ac-config-data\") pod \"3e6332da-21f3-4136-94e9-c5302ba865ac\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.524727 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snn8f\" (UniqueName: \"kubernetes.io/projected/3e6332da-21f3-4136-94e9-c5302ba865ac-kube-api-access-snn8f\") pod \"3e6332da-21f3-4136-94e9-c5302ba865ac\" (UID: \"3e6332da-21f3-4136-94e9-c5302ba865ac\") " Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.540135 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e6332da-21f3-4136-94e9-c5302ba865ac-kube-api-access-snn8f" (OuterVolumeSpecName: "kube-api-access-snn8f") pod "3e6332da-21f3-4136-94e9-c5302ba865ac" (UID: "3e6332da-21f3-4136-94e9-c5302ba865ac"). InnerVolumeSpecName "kube-api-access-snn8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.589285 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6332da-21f3-4136-94e9-c5302ba865ac-config-data" (OuterVolumeSpecName: "config-data") pod "3e6332da-21f3-4136-94e9-c5302ba865ac" (UID: "3e6332da-21f3-4136-94e9-c5302ba865ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.627677 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e6332da-21f3-4136-94e9-c5302ba865ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.627724 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snn8f\" (UniqueName: \"kubernetes.io/projected/3e6332da-21f3-4136-94e9-c5302ba865ac-kube-api-access-snn8f\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.763556 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" event={"ID":"3e6332da-21f3-4136-94e9-c5302ba865ac","Type":"ContainerDied","Data":"4e14d5beaeb2dccf8e1c99fb397873c94989d1dcea598c79379a9993a6c129e3"} Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.764096 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e14d5beaeb2dccf8e1c99fb397873c94989d1dcea598c79379a9993a6c129e3" Oct 07 12:42:15 crc kubenswrapper[4702]: I1007 12:42:15.763620 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-8dlbs" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.734612 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-fxd5x"] Oct 07 12:42:16 crc kubenswrapper[4702]: E1007 12:42:16.735060 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e6332da-21f3-4136-94e9-c5302ba865ac" containerName="keystone-db-sync" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.735082 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e6332da-21f3-4136-94e9-c5302ba865ac" containerName="keystone-db-sync" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.735249 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e6332da-21f3-4136-94e9-c5302ba865ac" containerName="keystone-db-sync" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.738860 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.741496 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-bdjdn" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.741734 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.741920 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.742170 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.747248 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-fxd5x"] Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.845832 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-credential-keys\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.845879 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-config-data\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.846056 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbr8h\" (UniqueName: \"kubernetes.io/projected/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-kube-api-access-dbr8h\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.846163 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-scripts\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.846190 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-fernet-keys\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.947934 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-scripts\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.948040 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-fernet-keys\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.948086 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-credential-keys\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.948133 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-config-data\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.948214 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbr8h\" (UniqueName: \"kubernetes.io/projected/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-kube-api-access-dbr8h\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.955863 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-fernet-keys\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.956278 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-scripts\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.956453 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-config-data\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.959339 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-credential-keys\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:16 crc kubenswrapper[4702]: I1007 12:42:16.970203 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbr8h\" (UniqueName: \"kubernetes.io/projected/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-kube-api-access-dbr8h\") pod \"keystone-bootstrap-fxd5x\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.069211 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.193033 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-lg4mj"] Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.557795 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-fxd5x"] Oct 07 12:42:17 crc kubenswrapper[4702]: W1007 12:42:17.568213 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod749ab4c9_2ed8_4e0d_b905_bd8e6719d322.slice/crio-05c6139911b4f5ae72a0eda3ff4a6db4e5b4e2aa5c950a3cae2e2fdd424aea7d WatchSource:0}: Error finding container 05c6139911b4f5ae72a0eda3ff4a6db4e5b4e2aa5c950a3cae2e2fdd424aea7d: Status 404 returned error can't find the container with id 05c6139911b4f5ae72a0eda3ff4a6db4e5b4e2aa5c950a3cae2e2fdd424aea7d Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.785577 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-lg4mj" event={"ID":"63310df2-8f1e-4ddc-b41f-f338b754812e","Type":"ContainerStarted","Data":"96bb2045448f1d298063063beaa698a1ab797168aab318aac3519cc2ea0626d0"} Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.787804 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-r4dwv" event={"ID":"bde85707-06b8-407a-9f37-51d6be9c666e","Type":"ContainerStarted","Data":"afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4"} Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.787975 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-r4dwv" podUID="bde85707-06b8-407a-9f37-51d6be9c666e" containerName="registry-server" containerID="cri-o://afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4" gracePeriod=2 Oct 07 12:42:17 crc kubenswrapper[4702]: I1007 12:42:17.789656 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" event={"ID":"749ab4c9-2ed8-4e0d-b905-bd8e6719d322","Type":"ContainerStarted","Data":"05c6139911b4f5ae72a0eda3ff4a6db4e5b4e2aa5c950a3cae2e2fdd424aea7d"} Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.626498 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.787885 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbv96\" (UniqueName: \"kubernetes.io/projected/bde85707-06b8-407a-9f37-51d6be9c666e-kube-api-access-gbv96\") pod \"bde85707-06b8-407a-9f37-51d6be9c666e\" (UID: \"bde85707-06b8-407a-9f37-51d6be9c666e\") " Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.793003 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde85707-06b8-407a-9f37-51d6be9c666e-kube-api-access-gbv96" (OuterVolumeSpecName: "kube-api-access-gbv96") pod "bde85707-06b8-407a-9f37-51d6be9c666e" (UID: "bde85707-06b8-407a-9f37-51d6be9c666e"). InnerVolumeSpecName "kube-api-access-gbv96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.803450 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" event={"ID":"749ab4c9-2ed8-4e0d-b905-bd8e6719d322","Type":"ContainerStarted","Data":"d2485110569ea9a25f6e03d3cfe7eab686bbd47024adaadfe40eab7e09526720"} Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.805880 4702 generic.go:334] "Generic (PLEG): container finished" podID="bde85707-06b8-407a-9f37-51d6be9c666e" containerID="afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4" exitCode=0 Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.805966 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-r4dwv" event={"ID":"bde85707-06b8-407a-9f37-51d6be9c666e","Type":"ContainerDied","Data":"afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4"} Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.806034 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-r4dwv" event={"ID":"bde85707-06b8-407a-9f37-51d6be9c666e","Type":"ContainerDied","Data":"99887b7e3e37aca29eda22873f4c86f6abacd5dcc082a3975345bf83b90c8c65"} Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.806054 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-r4dwv" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.806060 4702 scope.go:117] "RemoveContainer" containerID="afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.829990 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" podStartSLOduration=2.829948883 podStartE2EDuration="2.829948883s" podCreationTimestamp="2025-10-07 12:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:42:18.822433665 +0000 UTC m=+1094.998169821" watchObservedRunningTime="2025-10-07 12:42:18.829948883 +0000 UTC m=+1095.005684989" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.833685 4702 scope.go:117] "RemoveContainer" containerID="afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4" Oct 07 12:42:18 crc kubenswrapper[4702]: E1007 12:42:18.834331 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4\": container with ID starting with afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4 not found: ID does not exist" containerID="afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.834376 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4"} err="failed to get container status \"afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4\": rpc error: code = NotFound desc = could not find container \"afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4\": container with ID starting with afaad1f84cdc46a7a593e3920d046d9d0d95aec151e50863f6a4c05b4461dde4 not found: ID does not exist" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.889842 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbv96\" (UniqueName: \"kubernetes.io/projected/bde85707-06b8-407a-9f37-51d6be9c666e-kube-api-access-gbv96\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.894931 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-r4dwv"] Oct 07 12:42:18 crc kubenswrapper[4702]: I1007 12:42:18.903642 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-r4dwv"] Oct 07 12:42:19 crc kubenswrapper[4702]: I1007 12:42:19.820039 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-lg4mj" event={"ID":"63310df2-8f1e-4ddc-b41f-f338b754812e","Type":"ContainerStarted","Data":"40675a20ee567a210edc8036575f72d7ac42c5c655940e6298a7bede0f3c1164"} Oct 07 12:42:19 crc kubenswrapper[4702]: I1007 12:42:19.828539 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" event={"ID":"5d1fc69f-1999-40a0-acf6-eba215843669","Type":"ContainerStarted","Data":"1bc018067f35cad5d5fc90351e48aaccb4a35730e2763d9053750cbf616eb9ac"} Oct 07 12:42:19 crc kubenswrapper[4702]: I1007 12:42:19.843584 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-lg4mj" podStartSLOduration=5.387186678 podStartE2EDuration="6.843541597s" podCreationTimestamp="2025-10-07 12:42:13 +0000 UTC" firstStartedPulling="2025-10-07 12:42:17.214876251 +0000 UTC m=+1093.390612367" lastFinishedPulling="2025-10-07 12:42:18.67123115 +0000 UTC m=+1094.846967286" observedRunningTime="2025-10-07 12:42:19.843379553 +0000 UTC m=+1096.019115709" watchObservedRunningTime="2025-10-07 12:42:19.843541597 +0000 UTC m=+1096.019277753" Oct 07 12:42:19 crc kubenswrapper[4702]: I1007 12:42:19.885298 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" podStartSLOduration=2.284105095 podStartE2EDuration="12.885269245s" podCreationTimestamp="2025-10-07 12:42:07 +0000 UTC" firstStartedPulling="2025-10-07 12:42:08.067767164 +0000 UTC m=+1084.243503280" lastFinishedPulling="2025-10-07 12:42:18.668931314 +0000 UTC m=+1094.844667430" observedRunningTime="2025-10-07 12:42:19.878186 +0000 UTC m=+1096.053922116" watchObservedRunningTime="2025-10-07 12:42:19.885269245 +0000 UTC m=+1096.061005401" Oct 07 12:42:20 crc kubenswrapper[4702]: I1007 12:42:20.653163 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde85707-06b8-407a-9f37-51d6be9c666e" path="/var/lib/kubelet/pods/bde85707-06b8-407a-9f37-51d6be9c666e/volumes" Oct 07 12:42:21 crc kubenswrapper[4702]: I1007 12:42:21.862877 4702 generic.go:334] "Generic (PLEG): container finished" podID="749ab4c9-2ed8-4e0d-b905-bd8e6719d322" containerID="d2485110569ea9a25f6e03d3cfe7eab686bbd47024adaadfe40eab7e09526720" exitCode=0 Oct 07 12:42:21 crc kubenswrapper[4702]: I1007 12:42:21.863234 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" event={"ID":"749ab4c9-2ed8-4e0d-b905-bd8e6719d322","Type":"ContainerDied","Data":"d2485110569ea9a25f6e03d3cfe7eab686bbd47024adaadfe40eab7e09526720"} Oct 07 12:42:22 crc kubenswrapper[4702]: I1007 12:42:22.817230 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:22 crc kubenswrapper[4702]: E1007 12:42:22.817552 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 07 12:42:22 crc kubenswrapper[4702]: E1007 12:42:22.817572 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 07 12:42:22 crc kubenswrapper[4702]: E1007 12:42:22.817640 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift podName:7646503c-ea03-4ce3-9438-54b187e1aec3 nodeName:}" failed. No retries permitted until 2025-10-07 12:42:38.817615548 +0000 UTC m=+1114.993351674 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift") pod "swift-storage-0" (UID: "7646503c-ea03-4ce3-9438-54b187e1aec3") : configmap "swift-ring-files" not found Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.188096 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.223174 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-config-data\") pod \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.223259 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-fernet-keys\") pod \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.223320 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-scripts\") pod \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.223392 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbr8h\" (UniqueName: \"kubernetes.io/projected/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-kube-api-access-dbr8h\") pod \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.223436 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-credential-keys\") pod \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\" (UID: \"749ab4c9-2ed8-4e0d-b905-bd8e6719d322\") " Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.230269 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-scripts" (OuterVolumeSpecName: "scripts") pod "749ab4c9-2ed8-4e0d-b905-bd8e6719d322" (UID: "749ab4c9-2ed8-4e0d-b905-bd8e6719d322"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.230990 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "749ab4c9-2ed8-4e0d-b905-bd8e6719d322" (UID: "749ab4c9-2ed8-4e0d-b905-bd8e6719d322"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.232414 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "749ab4c9-2ed8-4e0d-b905-bd8e6719d322" (UID: "749ab4c9-2ed8-4e0d-b905-bd8e6719d322"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.232476 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-kube-api-access-dbr8h" (OuterVolumeSpecName: "kube-api-access-dbr8h") pod "749ab4c9-2ed8-4e0d-b905-bd8e6719d322" (UID: "749ab4c9-2ed8-4e0d-b905-bd8e6719d322"). InnerVolumeSpecName "kube-api-access-dbr8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.244167 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-config-data" (OuterVolumeSpecName: "config-data") pod "749ab4c9-2ed8-4e0d-b905-bd8e6719d322" (UID: "749ab4c9-2ed8-4e0d-b905-bd8e6719d322"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.325503 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.325544 4702 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.325554 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.325567 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbr8h\" (UniqueName: \"kubernetes.io/projected/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-kube-api-access-dbr8h\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.325577 4702 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/749ab4c9-2ed8-4e0d-b905-bd8e6719d322-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.812071 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.812161 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.845265 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.880069 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.880080 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-fxd5x" event={"ID":"749ab4c9-2ed8-4e0d-b905-bd8e6719d322","Type":"ContainerDied","Data":"05c6139911b4f5ae72a0eda3ff4a6db4e5b4e2aa5c950a3cae2e2fdd424aea7d"} Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.880685 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05c6139911b4f5ae72a0eda3ff4a6db4e5b4e2aa5c950a3cae2e2fdd424aea7d" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.987193 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-69787db686-hgjdk"] Oct 07 12:42:23 crc kubenswrapper[4702]: E1007 12:42:23.987552 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde85707-06b8-407a-9f37-51d6be9c666e" containerName="registry-server" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.987576 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde85707-06b8-407a-9f37-51d6be9c666e" containerName="registry-server" Oct 07 12:42:23 crc kubenswrapper[4702]: E1007 12:42:23.987606 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="749ab4c9-2ed8-4e0d-b905-bd8e6719d322" containerName="keystone-bootstrap" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.987615 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="749ab4c9-2ed8-4e0d-b905-bd8e6719d322" containerName="keystone-bootstrap" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.987763 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde85707-06b8-407a-9f37-51d6be9c666e" containerName="registry-server" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.987781 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="749ab4c9-2ed8-4e0d-b905-bd8e6719d322" containerName="keystone-bootstrap" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.988411 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.991106 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.991213 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.992037 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 07 12:42:23 crc kubenswrapper[4702]: I1007 12:42:23.992188 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-bdjdn" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.003624 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-69787db686-hgjdk"] Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.035432 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-credential-keys\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.035495 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-fernet-keys\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.035516 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-scripts\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.035538 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t58j\" (UniqueName: \"kubernetes.io/projected/95471c20-4000-4f65-9f81-255c1b909f35-kube-api-access-6t58j\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.035568 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-config-data\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.136216 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-credential-keys\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.136278 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-fernet-keys\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.136319 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-scripts\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.136344 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t58j\" (UniqueName: \"kubernetes.io/projected/95471c20-4000-4f65-9f81-255c1b909f35-kube-api-access-6t58j\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.136376 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-config-data\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.142398 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-scripts\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.142534 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-credential-keys\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.143841 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-config-data\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.144512 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95471c20-4000-4f65-9f81-255c1b909f35-fernet-keys\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.160705 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t58j\" (UniqueName: \"kubernetes.io/projected/95471c20-4000-4f65-9f81-255c1b909f35-kube-api-access-6t58j\") pod \"keystone-69787db686-hgjdk\" (UID: \"95471c20-4000-4f65-9f81-255c1b909f35\") " pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.306563 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.742752 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-69787db686-hgjdk"] Oct 07 12:42:24 crc kubenswrapper[4702]: W1007 12:42:24.755109 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95471c20_4000_4f65_9f81_255c1b909f35.slice/crio-f3b36b5d202f11c1c2b4a8d37a7bde0bfb2509bddf42ae313501e90518c60bcc WatchSource:0}: Error finding container f3b36b5d202f11c1c2b4a8d37a7bde0bfb2509bddf42ae313501e90518c60bcc: Status 404 returned error can't find the container with id f3b36b5d202f11c1c2b4a8d37a7bde0bfb2509bddf42ae313501e90518c60bcc Oct 07 12:42:24 crc kubenswrapper[4702]: I1007 12:42:24.888500 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" event={"ID":"95471c20-4000-4f65-9f81-255c1b909f35","Type":"ContainerStarted","Data":"f3b36b5d202f11c1c2b4a8d37a7bde0bfb2509bddf42ae313501e90518c60bcc"} Oct 07 12:42:25 crc kubenswrapper[4702]: I1007 12:42:25.899266 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" event={"ID":"95471c20-4000-4f65-9f81-255c1b909f35","Type":"ContainerStarted","Data":"2d933f00215369caad60e273170c48c387bcbdd1e902164164c561fc266e23ca"} Oct 07 12:42:25 crc kubenswrapper[4702]: I1007 12:42:25.900315 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:42:25 crc kubenswrapper[4702]: I1007 12:42:25.902711 4702 generic.go:334] "Generic (PLEG): container finished" podID="5d1fc69f-1999-40a0-acf6-eba215843669" containerID="1bc018067f35cad5d5fc90351e48aaccb4a35730e2763d9053750cbf616eb9ac" exitCode=0 Oct 07 12:42:25 crc kubenswrapper[4702]: I1007 12:42:25.902805 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" event={"ID":"5d1fc69f-1999-40a0-acf6-eba215843669","Type":"ContainerDied","Data":"1bc018067f35cad5d5fc90351e48aaccb4a35730e2763d9053750cbf616eb9ac"} Oct 07 12:42:25 crc kubenswrapper[4702]: I1007 12:42:25.936146 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" podStartSLOduration=2.93610207 podStartE2EDuration="2.93610207s" podCreationTimestamp="2025-10-07 12:42:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:42:25.927595224 +0000 UTC m=+1102.103331350" watchObservedRunningTime="2025-10-07 12:42:25.93610207 +0000 UTC m=+1102.111838256" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.281229 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.296971 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-scripts\") pod \"5d1fc69f-1999-40a0-acf6-eba215843669\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.297160 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-dispersionconf\") pod \"5d1fc69f-1999-40a0-acf6-eba215843669\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.297340 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-ring-data-devices\") pod \"5d1fc69f-1999-40a0-acf6-eba215843669\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.297493 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54cvr\" (UniqueName: \"kubernetes.io/projected/5d1fc69f-1999-40a0-acf6-eba215843669-kube-api-access-54cvr\") pod \"5d1fc69f-1999-40a0-acf6-eba215843669\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.297578 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d1fc69f-1999-40a0-acf6-eba215843669-etc-swift\") pod \"5d1fc69f-1999-40a0-acf6-eba215843669\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.297660 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-swiftconf\") pod \"5d1fc69f-1999-40a0-acf6-eba215843669\" (UID: \"5d1fc69f-1999-40a0-acf6-eba215843669\") " Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.299481 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5d1fc69f-1999-40a0-acf6-eba215843669" (UID: "5d1fc69f-1999-40a0-acf6-eba215843669"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.300531 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d1fc69f-1999-40a0-acf6-eba215843669-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5d1fc69f-1999-40a0-acf6-eba215843669" (UID: "5d1fc69f-1999-40a0-acf6-eba215843669"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.301019 4702 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.301055 4702 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d1fc69f-1999-40a0-acf6-eba215843669-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.308243 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5d1fc69f-1999-40a0-acf6-eba215843669" (UID: "5d1fc69f-1999-40a0-acf6-eba215843669"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.313181 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d1fc69f-1999-40a0-acf6-eba215843669-kube-api-access-54cvr" (OuterVolumeSpecName: "kube-api-access-54cvr") pod "5d1fc69f-1999-40a0-acf6-eba215843669" (UID: "5d1fc69f-1999-40a0-acf6-eba215843669"). InnerVolumeSpecName "kube-api-access-54cvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.321638 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5d1fc69f-1999-40a0-acf6-eba215843669" (UID: "5d1fc69f-1999-40a0-acf6-eba215843669"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.324428 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-scripts" (OuterVolumeSpecName: "scripts") pod "5d1fc69f-1999-40a0-acf6-eba215843669" (UID: "5d1fc69f-1999-40a0-acf6-eba215843669"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.402183 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d1fc69f-1999-40a0-acf6-eba215843669-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.402548 4702 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.402622 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54cvr\" (UniqueName: \"kubernetes.io/projected/5d1fc69f-1999-40a0-acf6-eba215843669-kube-api-access-54cvr\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.402680 4702 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d1fc69f-1999-40a0-acf6-eba215843669-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.922184 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" event={"ID":"5d1fc69f-1999-40a0-acf6-eba215843669","Type":"ContainerDied","Data":"b9e61485775b9102b82e5e558543ed154ac6d17b0b04eccb8dafabbb335c7810"} Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.922239 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9e61485775b9102b82e5e558543ed154ac6d17b0b04eccb8dafabbb335c7810" Oct 07 12:42:27 crc kubenswrapper[4702]: I1007 12:42:27.922374 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-gv9w7" Oct 07 12:42:33 crc kubenswrapper[4702]: I1007 12:42:33.845415 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-lg4mj" Oct 07 12:42:38 crc kubenswrapper[4702]: I1007 12:42:38.897624 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:38 crc kubenswrapper[4702]: I1007 12:42:38.908459 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7646503c-ea03-4ce3-9438-54b187e1aec3-etc-swift\") pod \"swift-storage-0\" (UID: \"7646503c-ea03-4ce3-9438-54b187e1aec3\") " pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:39 crc kubenswrapper[4702]: I1007 12:42:39.068918 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 07 12:42:39 crc kubenswrapper[4702]: I1007 12:42:39.263851 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:42:39 crc kubenswrapper[4702]: I1007 12:42:39.264892 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:42:39 crc kubenswrapper[4702]: I1007 12:42:39.351682 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 07 12:42:40 crc kubenswrapper[4702]: I1007 12:42:40.025941 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"b64538a3762f0596458485fb2219457218992ba427abbedef0e53bd50d5b91e5"} Oct 07 12:42:41 crc kubenswrapper[4702]: I1007 12:42:41.039236 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"50ae72482e1af0a56540cbab7cd443ddf7e0d53a1fbb804be57de49e5feb197e"} Oct 07 12:42:41 crc kubenswrapper[4702]: I1007 12:42:41.039657 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"38cc33696edb2ad64f26204b4a715e9b1cb4675a08a2020a9c5331ac6779899e"} Oct 07 12:42:41 crc kubenswrapper[4702]: I1007 12:42:41.039667 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"6d4647cb5e1448afd45c32561300afc2247f5db503bba33fb3bb30606ae262cb"} Oct 07 12:42:42 crc kubenswrapper[4702]: I1007 12:42:42.055614 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"e40c24d0b7e5978cc9924a4284b0c86a726a039ca8b9e11a486a941897802742"} Oct 07 12:42:43 crc kubenswrapper[4702]: I1007 12:42:43.085045 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"9a024ca4aa06b5c758199567bb0525d9e2760e1c874644755de09ac23720b6a1"} Oct 07 12:42:43 crc kubenswrapper[4702]: I1007 12:42:43.085656 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"3df221c63cf85b5195f90a4e1771a00564265b8c1d7f64225b9431d495990f89"} Oct 07 12:42:43 crc kubenswrapper[4702]: I1007 12:42:43.085687 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"f9f7fefd1ea1db057cb529dcf83f01dd28ce50e40a09c5ffd31cc78453fb746f"} Oct 07 12:42:43 crc kubenswrapper[4702]: I1007 12:42:43.085704 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"f9f788871a29bf980bb0c866c31d85368f3c612fc593f57b678436ad02de76f1"} Oct 07 12:42:44 crc kubenswrapper[4702]: I1007 12:42:44.109613 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"2285a135022725909805e0c82cdce7a0b89166ae13259c1296997bb5c63c2739"} Oct 07 12:42:44 crc kubenswrapper[4702]: I1007 12:42:44.110076 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"4e5f6a471f3d2c26b9481cb0416c528ebd79b6515443020cd476e22a8887b4fb"} Oct 07 12:42:45 crc kubenswrapper[4702]: I1007 12:42:45.128807 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"47a63acd81cf21b1051ee6b3b56c4b0f4cc3cb7aa72061d636030449cd23ccb9"} Oct 07 12:42:45 crc kubenswrapper[4702]: I1007 12:42:45.129287 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"4deae4cba3a7e0f7309c7abb7c89d199ade6ae8f3597728dd9d8fac4d2f19427"} Oct 07 12:42:45 crc kubenswrapper[4702]: I1007 12:42:45.129304 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"78856519adea33b9a202e50b8e8b966f37ed3180c645b0839864800969fa21da"} Oct 07 12:42:45 crc kubenswrapper[4702]: I1007 12:42:45.129318 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"d9bb6fa2e818ec885ef7b3845c313071c53af0c6e98b3fb052529cf1b5e18236"} Oct 07 12:42:45 crc kubenswrapper[4702]: I1007 12:42:45.129329 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7646503c-ea03-4ce3-9438-54b187e1aec3","Type":"ContainerStarted","Data":"cdecf9941f33919a2d552690f7ec3affaada5fb97e70c12c7efbd20836ebfd79"} Oct 07 12:42:45 crc kubenswrapper[4702]: I1007 12:42:45.175185 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=35.846548776 podStartE2EDuration="40.175152217s" podCreationTimestamp="2025-10-07 12:42:05 +0000 UTC" firstStartedPulling="2025-10-07 12:42:39.371780196 +0000 UTC m=+1115.547516312" lastFinishedPulling="2025-10-07 12:42:43.700383637 +0000 UTC m=+1119.876119753" observedRunningTime="2025-10-07 12:42:45.161150629 +0000 UTC m=+1121.336886785" watchObservedRunningTime="2025-10-07 12:42:45.175152217 +0000 UTC m=+1121.350888333" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.368717 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4"] Oct 07 12:42:46 crc kubenswrapper[4702]: E1007 12:42:46.369707 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1fc69f-1999-40a0-acf6-eba215843669" containerName="swift-ring-rebalance" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.369729 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1fc69f-1999-40a0-acf6-eba215843669" containerName="swift-ring-rebalance" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.369947 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d1fc69f-1999-40a0-acf6-eba215843669" containerName="swift-ring-rebalance" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.371356 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.402115 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrk4d" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.408597 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4"] Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.528012 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crqmm\" (UniqueName: \"kubernetes.io/projected/ac694397-13ed-4daa-9a05-7ae908c3da4f-kube-api-access-crqmm\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.528093 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-util\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.528196 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-bundle\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.630220 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crqmm\" (UniqueName: \"kubernetes.io/projected/ac694397-13ed-4daa-9a05-7ae908c3da4f-kube-api-access-crqmm\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.630301 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-util\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.630390 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-bundle\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.631328 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-bundle\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.631617 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-util\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.668784 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crqmm\" (UniqueName: \"kubernetes.io/projected/ac694397-13ed-4daa-9a05-7ae908c3da4f-kube-api-access-crqmm\") pod \"1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.727871 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:46 crc kubenswrapper[4702]: I1007 12:42:46.988411 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4"] Oct 07 12:42:47 crc kubenswrapper[4702]: I1007 12:42:47.152359 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" event={"ID":"ac694397-13ed-4daa-9a05-7ae908c3da4f","Type":"ContainerStarted","Data":"67150181c2fe6bbc895f91396e73bbb9b67e2771e2a39f1096c881f5a3919c85"} Oct 07 12:42:48 crc kubenswrapper[4702]: I1007 12:42:48.160467 4702 generic.go:334] "Generic (PLEG): container finished" podID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerID="3afd38fae9d98d76ac62798080ee8fb323b3f226b4a1282dd8c525b235884a08" exitCode=0 Oct 07 12:42:48 crc kubenswrapper[4702]: I1007 12:42:48.160568 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" event={"ID":"ac694397-13ed-4daa-9a05-7ae908c3da4f","Type":"ContainerDied","Data":"3afd38fae9d98d76ac62798080ee8fb323b3f226b4a1282dd8c525b235884a08"} Oct 07 12:42:49 crc kubenswrapper[4702]: I1007 12:42:49.175609 4702 generic.go:334] "Generic (PLEG): container finished" podID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerID="aff48b6001f3b0be6002fcc4c301eaf1f40720de806fb2ac08d00dbfacb034fc" exitCode=0 Oct 07 12:42:49 crc kubenswrapper[4702]: I1007 12:42:49.175781 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" event={"ID":"ac694397-13ed-4daa-9a05-7ae908c3da4f","Type":"ContainerDied","Data":"aff48b6001f3b0be6002fcc4c301eaf1f40720de806fb2ac08d00dbfacb034fc"} Oct 07 12:42:50 crc kubenswrapper[4702]: I1007 12:42:50.203563 4702 generic.go:334] "Generic (PLEG): container finished" podID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerID="369279980ec019b8259f80a42047c0bb634b62d7433e573f5893b8340acb337f" exitCode=0 Oct 07 12:42:50 crc kubenswrapper[4702]: I1007 12:42:50.203636 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" event={"ID":"ac694397-13ed-4daa-9a05-7ae908c3da4f","Type":"ContainerDied","Data":"369279980ec019b8259f80a42047c0bb634b62d7433e573f5893b8340acb337f"} Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.547451 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.707806 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-util\") pod \"ac694397-13ed-4daa-9a05-7ae908c3da4f\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.708455 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crqmm\" (UniqueName: \"kubernetes.io/projected/ac694397-13ed-4daa-9a05-7ae908c3da4f-kube-api-access-crqmm\") pod \"ac694397-13ed-4daa-9a05-7ae908c3da4f\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.708570 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-bundle\") pod \"ac694397-13ed-4daa-9a05-7ae908c3da4f\" (UID: \"ac694397-13ed-4daa-9a05-7ae908c3da4f\") " Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.710462 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-bundle" (OuterVolumeSpecName: "bundle") pod "ac694397-13ed-4daa-9a05-7ae908c3da4f" (UID: "ac694397-13ed-4daa-9a05-7ae908c3da4f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.716825 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac694397-13ed-4daa-9a05-7ae908c3da4f-kube-api-access-crqmm" (OuterVolumeSpecName: "kube-api-access-crqmm") pod "ac694397-13ed-4daa-9a05-7ae908c3da4f" (UID: "ac694397-13ed-4daa-9a05-7ae908c3da4f"). InnerVolumeSpecName "kube-api-access-crqmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.732110 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-util" (OuterVolumeSpecName: "util") pod "ac694397-13ed-4daa-9a05-7ae908c3da4f" (UID: "ac694397-13ed-4daa-9a05-7ae908c3da4f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.811454 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-util\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.812358 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crqmm\" (UniqueName: \"kubernetes.io/projected/ac694397-13ed-4daa-9a05-7ae908c3da4f-kube-api-access-crqmm\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:51 crc kubenswrapper[4702]: I1007 12:42:51.812397 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac694397-13ed-4daa-9a05-7ae908c3da4f-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:42:52 crc kubenswrapper[4702]: I1007 12:42:52.234556 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" event={"ID":"ac694397-13ed-4daa-9a05-7ae908c3da4f","Type":"ContainerDied","Data":"67150181c2fe6bbc895f91396e73bbb9b67e2771e2a39f1096c881f5a3919c85"} Oct 07 12:42:52 crc kubenswrapper[4702]: I1007 12:42:52.234628 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67150181c2fe6bbc895f91396e73bbb9b67e2771e2a39f1096c881f5a3919c85" Oct 07 12:42:52 crc kubenswrapper[4702]: I1007 12:42:52.234741 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4" Oct 07 12:42:55 crc kubenswrapper[4702]: I1007 12:42:55.736052 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-69787db686-hgjdk" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.789568 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7"] Oct 07 12:43:01 crc kubenswrapper[4702]: E1007 12:43:01.790497 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="util" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.790512 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="util" Oct 07 12:43:01 crc kubenswrapper[4702]: E1007 12:43:01.790543 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="pull" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.790552 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="pull" Oct 07 12:43:01 crc kubenswrapper[4702]: E1007 12:43:01.790562 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="extract" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.790568 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="extract" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.790685 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac694397-13ed-4daa-9a05-7ae908c3da4f" containerName="extract" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.791424 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.795103 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-md85w" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.799315 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.804415 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7"] Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.877892 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srftm\" (UniqueName: \"kubernetes.io/projected/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-kube-api-access-srftm\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.877996 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-webhook-cert\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.878056 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-apiservice-cert\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.979093 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srftm\" (UniqueName: \"kubernetes.io/projected/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-kube-api-access-srftm\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.979205 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-webhook-cert\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.980245 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-apiservice-cert\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.986444 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-apiservice-cert\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.997009 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-webhook-cert\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:01 crc kubenswrapper[4702]: I1007 12:43:01.998428 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srftm\" (UniqueName: \"kubernetes.io/projected/fac7e760-a6e5-42dd-8cdd-ebcf25bba335-kube-api-access-srftm\") pod \"glance-operator-controller-manager-85f8db96c5-w6ds7\" (UID: \"fac7e760-a6e5-42dd-8cdd-ebcf25bba335\") " pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:02 crc kubenswrapper[4702]: I1007 12:43:02.124199 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:02 crc kubenswrapper[4702]: I1007 12:43:02.675847 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7"] Oct 07 12:43:02 crc kubenswrapper[4702]: W1007 12:43:02.685411 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfac7e760_a6e5_42dd_8cdd_ebcf25bba335.slice/crio-23cd80ef17cb2803998b30aa2211ac1fc8bb01e6b7da71b1736fd9dfd547f394 WatchSource:0}: Error finding container 23cd80ef17cb2803998b30aa2211ac1fc8bb01e6b7da71b1736fd9dfd547f394: Status 404 returned error can't find the container with id 23cd80ef17cb2803998b30aa2211ac1fc8bb01e6b7da71b1736fd9dfd547f394 Oct 07 12:43:03 crc kubenswrapper[4702]: I1007 12:43:03.323364 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" event={"ID":"fac7e760-a6e5-42dd-8cdd-ebcf25bba335","Type":"ContainerStarted","Data":"23cd80ef17cb2803998b30aa2211ac1fc8bb01e6b7da71b1736fd9dfd547f394"} Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.222420 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn"] Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.224383 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.227318 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.243706 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn"] Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.334174 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8sxf\" (UniqueName: \"kubernetes.io/projected/0b851329-4872-417e-84c5-24b567021d51-kube-api-access-q8sxf\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.334275 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b851329-4872-417e-84c5-24b567021d51-config-data\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.334336 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b851329-4872-417e-84c5-24b567021d51-log-httpd\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.334377 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b851329-4872-417e-84c5-24b567021d51-run-httpd\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.334440 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0b851329-4872-417e-84c5-24b567021d51-etc-swift\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.435683 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b851329-4872-417e-84c5-24b567021d51-log-httpd\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.436159 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b851329-4872-417e-84c5-24b567021d51-run-httpd\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.436296 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0b851329-4872-417e-84c5-24b567021d51-etc-swift\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.436406 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b851329-4872-417e-84c5-24b567021d51-log-httpd\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.436542 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8sxf\" (UniqueName: \"kubernetes.io/projected/0b851329-4872-417e-84c5-24b567021d51-kube-api-access-q8sxf\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.436655 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b851329-4872-417e-84c5-24b567021d51-config-data\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.436936 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b851329-4872-417e-84c5-24b567021d51-run-httpd\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.444460 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0b851329-4872-417e-84c5-24b567021d51-etc-swift\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.457977 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8sxf\" (UniqueName: \"kubernetes.io/projected/0b851329-4872-417e-84c5-24b567021d51-kube-api-access-q8sxf\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.459745 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b851329-4872-417e-84c5-24b567021d51-config-data\") pod \"swift-proxy-6dd8f59749-9f2cn\" (UID: \"0b851329-4872-417e-84c5-24b567021d51\") " pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.544605 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:05 crc kubenswrapper[4702]: I1007 12:43:05.792567 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn"] Oct 07 12:43:06 crc kubenswrapper[4702]: W1007 12:43:06.072066 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b851329_4872_417e_84c5_24b567021d51.slice/crio-4f2ef40b3e522e8cffbc8854a43049c867045725fda97d9e32c029040c725727 WatchSource:0}: Error finding container 4f2ef40b3e522e8cffbc8854a43049c867045725fda97d9e32c029040c725727: Status 404 returned error can't find the container with id 4f2ef40b3e522e8cffbc8854a43049c867045725fda97d9e32c029040c725727 Oct 07 12:43:06 crc kubenswrapper[4702]: I1007 12:43:06.346703 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" event={"ID":"0b851329-4872-417e-84c5-24b567021d51","Type":"ContainerStarted","Data":"c7307791f73323ce84da73e41a0fb781d6b801f4c3363f60cfabed444623f04f"} Oct 07 12:43:06 crc kubenswrapper[4702]: I1007 12:43:06.347282 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" event={"ID":"0b851329-4872-417e-84c5-24b567021d51","Type":"ContainerStarted","Data":"4f2ef40b3e522e8cffbc8854a43049c867045725fda97d9e32c029040c725727"} Oct 07 12:43:06 crc kubenswrapper[4702]: I1007 12:43:06.349155 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" event={"ID":"fac7e760-a6e5-42dd-8cdd-ebcf25bba335","Type":"ContainerStarted","Data":"97be4c86519fecbd82a2f601a010358b82f97a36d8b62d581257de13be86c159"} Oct 07 12:43:07 crc kubenswrapper[4702]: I1007 12:43:07.362575 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" event={"ID":"0b851329-4872-417e-84c5-24b567021d51","Type":"ContainerStarted","Data":"3f4d8131047b0e888ffe8e67e30fb6101137ca1ae83304cf9694c27c35ba146e"} Oct 07 12:43:07 crc kubenswrapper[4702]: I1007 12:43:07.363163 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:07 crc kubenswrapper[4702]: I1007 12:43:07.363181 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:07 crc kubenswrapper[4702]: I1007 12:43:07.400979 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" podStartSLOduration=2.400942995 podStartE2EDuration="2.400942995s" podCreationTimestamp="2025-10-07 12:43:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:43:07.388507113 +0000 UTC m=+1143.564243259" watchObservedRunningTime="2025-10-07 12:43:07.400942995 +0000 UTC m=+1143.576679161" Oct 07 12:43:08 crc kubenswrapper[4702]: I1007 12:43:08.372443 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" event={"ID":"fac7e760-a6e5-42dd-8cdd-ebcf25bba335","Type":"ContainerStarted","Data":"afa95e905c264fe911153e5adb8b8c240e3012a1bdedf625821c963750ceda97"} Oct 07 12:43:08 crc kubenswrapper[4702]: I1007 12:43:08.397615 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" podStartSLOduration=2.785987132 podStartE2EDuration="7.397579815s" podCreationTimestamp="2025-10-07 12:43:01 +0000 UTC" firstStartedPulling="2025-10-07 12:43:02.689366857 +0000 UTC m=+1138.865102973" lastFinishedPulling="2025-10-07 12:43:07.30095954 +0000 UTC m=+1143.476695656" observedRunningTime="2025-10-07 12:43:08.393949029 +0000 UTC m=+1144.569685185" watchObservedRunningTime="2025-10-07 12:43:08.397579815 +0000 UTC m=+1144.573315941" Oct 07 12:43:09 crc kubenswrapper[4702]: I1007 12:43:09.263110 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:43:09 crc kubenswrapper[4702]: I1007 12:43:09.263710 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:43:09 crc kubenswrapper[4702]: I1007 12:43:09.263776 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:43:09 crc kubenswrapper[4702]: I1007 12:43:09.264654 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03f130f70c653ad000f84ff22f5df3f9d3bfff20eccc689a18ef25482e6731a7"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:43:09 crc kubenswrapper[4702]: I1007 12:43:09.264736 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://03f130f70c653ad000f84ff22f5df3f9d3bfff20eccc689a18ef25482e6731a7" gracePeriod=600 Oct 07 12:43:09 crc kubenswrapper[4702]: I1007 12:43:09.382329 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:10 crc kubenswrapper[4702]: I1007 12:43:10.392551 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="03f130f70c653ad000f84ff22f5df3f9d3bfff20eccc689a18ef25482e6731a7" exitCode=0 Oct 07 12:43:10 crc kubenswrapper[4702]: I1007 12:43:10.392634 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"03f130f70c653ad000f84ff22f5df3f9d3bfff20eccc689a18ef25482e6731a7"} Oct 07 12:43:10 crc kubenswrapper[4702]: I1007 12:43:10.392714 4702 scope.go:117] "RemoveContainer" containerID="3ca59d66613e781014a1bca3d21d490470064dd363a5feb6c9d2c05e7748d68e" Oct 07 12:43:11 crc kubenswrapper[4702]: I1007 12:43:11.400965 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"f8baadc91bcb473e7996507e561b2cb8a6841ec268f00c5e69f87373e5816c4b"} Oct 07 12:43:12 crc kubenswrapper[4702]: I1007 12:43:12.134060 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-85f8db96c5-w6ds7" Oct 07 12:43:15 crc kubenswrapper[4702]: I1007 12:43:15.547199 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:15 crc kubenswrapper[4702]: I1007 12:43:15.548091 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6dd8f59749-9f2cn" Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.930178 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.931971 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.936166 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.936471 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-kf8kh" Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.936681 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.941694 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 07 12:43:16 crc kubenswrapper[4702]: I1007 12:43:16.946284 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.007994 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.008072 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgm57\" (UniqueName: \"kubernetes.io/projected/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-kube-api-access-zgm57\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.008227 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-scripts\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.008524 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config-secret\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.110419 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.110808 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgm57\" (UniqueName: \"kubernetes.io/projected/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-kube-api-access-zgm57\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.110832 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-scripts\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.110871 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config-secret\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.111716 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.112485 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-scripts\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.128740 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config-secret\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.139597 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgm57\" (UniqueName: \"kubernetes.io/projected/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-kube-api-access-zgm57\") pod \"openstackclient\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.260159 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 07 12:43:17 crc kubenswrapper[4702]: I1007 12:43:17.537889 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:43:17 crc kubenswrapper[4702]: W1007 12:43:17.542868 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod766bd860_d6e5_4d16_8b5d_3825ec4dd9fe.slice/crio-234e6350bf57ac21a329a4e9c4bd1fa71d5281518ce6394fe4c37b06f8780723 WatchSource:0}: Error finding container 234e6350bf57ac21a329a4e9c4bd1fa71d5281518ce6394fe4c37b06f8780723: Status 404 returned error can't find the container with id 234e6350bf57ac21a329a4e9c4bd1fa71d5281518ce6394fe4c37b06f8780723 Oct 07 12:43:18 crc kubenswrapper[4702]: I1007 12:43:18.467512 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe","Type":"ContainerStarted","Data":"234e6350bf57ac21a329a4e9c4bd1fa71d5281518ce6394fe4c37b06f8780723"} Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.588982 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-2rqv8"] Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.592235 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.603624 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-2rqv8"] Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.684088 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nltqx\" (UniqueName: \"kubernetes.io/projected/2eaff209-6059-460b-86ed-a577b4abe041-kube-api-access-nltqx\") pod \"glance-db-create-2rqv8\" (UID: \"2eaff209-6059-460b-86ed-a577b4abe041\") " pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.785875 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nltqx\" (UniqueName: \"kubernetes.io/projected/2eaff209-6059-460b-86ed-a577b4abe041-kube-api-access-nltqx\") pod \"glance-db-create-2rqv8\" (UID: \"2eaff209-6059-460b-86ed-a577b4abe041\") " pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.825374 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nltqx\" (UniqueName: \"kubernetes.io/projected/2eaff209-6059-460b-86ed-a577b4abe041-kube-api-access-nltqx\") pod \"glance-db-create-2rqv8\" (UID: \"2eaff209-6059-460b-86ed-a577b4abe041\") " pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:21 crc kubenswrapper[4702]: I1007 12:43:21.922951 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:26 crc kubenswrapper[4702]: I1007 12:43:26.195396 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-2rqv8"] Oct 07 12:43:26 crc kubenswrapper[4702]: I1007 12:43:26.562966 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe","Type":"ContainerStarted","Data":"8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5"} Oct 07 12:43:26 crc kubenswrapper[4702]: I1007 12:43:26.564998 4702 generic.go:334] "Generic (PLEG): container finished" podID="2eaff209-6059-460b-86ed-a577b4abe041" containerID="5d4efdb54fbe55f8540e6cddc57ce58e31d4873d8d4320bef170963ce5793520" exitCode=0 Oct 07 12:43:26 crc kubenswrapper[4702]: I1007 12:43:26.565068 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-2rqv8" event={"ID":"2eaff209-6059-460b-86ed-a577b4abe041","Type":"ContainerDied","Data":"5d4efdb54fbe55f8540e6cddc57ce58e31d4873d8d4320bef170963ce5793520"} Oct 07 12:43:26 crc kubenswrapper[4702]: I1007 12:43:26.565126 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-2rqv8" event={"ID":"2eaff209-6059-460b-86ed-a577b4abe041","Type":"ContainerStarted","Data":"28ce8239b71f08706565ddade636eafe11b67981f199a876e4a97f342ab4fd41"} Oct 07 12:43:26 crc kubenswrapper[4702]: I1007 12:43:26.610961 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.282480354 podStartE2EDuration="10.610926821s" podCreationTimestamp="2025-10-07 12:43:16 +0000 UTC" firstStartedPulling="2025-10-07 12:43:17.545207727 +0000 UTC m=+1153.720943843" lastFinishedPulling="2025-10-07 12:43:25.873654184 +0000 UTC m=+1162.049390310" observedRunningTime="2025-10-07 12:43:26.581615297 +0000 UTC m=+1162.757351453" watchObservedRunningTime="2025-10-07 12:43:26.610926821 +0000 UTC m=+1162.786662977" Oct 07 12:43:27 crc kubenswrapper[4702]: I1007 12:43:27.873239 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:28 crc kubenswrapper[4702]: I1007 12:43:28.006970 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nltqx\" (UniqueName: \"kubernetes.io/projected/2eaff209-6059-460b-86ed-a577b4abe041-kube-api-access-nltqx\") pod \"2eaff209-6059-460b-86ed-a577b4abe041\" (UID: \"2eaff209-6059-460b-86ed-a577b4abe041\") " Oct 07 12:43:28 crc kubenswrapper[4702]: I1007 12:43:28.014455 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eaff209-6059-460b-86ed-a577b4abe041-kube-api-access-nltqx" (OuterVolumeSpecName: "kube-api-access-nltqx") pod "2eaff209-6059-460b-86ed-a577b4abe041" (UID: "2eaff209-6059-460b-86ed-a577b4abe041"). InnerVolumeSpecName "kube-api-access-nltqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:43:28 crc kubenswrapper[4702]: I1007 12:43:28.108390 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nltqx\" (UniqueName: \"kubernetes.io/projected/2eaff209-6059-460b-86ed-a577b4abe041-kube-api-access-nltqx\") on node \"crc\" DevicePath \"\"" Oct 07 12:43:28 crc kubenswrapper[4702]: I1007 12:43:28.587146 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-2rqv8" event={"ID":"2eaff209-6059-460b-86ed-a577b4abe041","Type":"ContainerDied","Data":"28ce8239b71f08706565ddade636eafe11b67981f199a876e4a97f342ab4fd41"} Oct 07 12:43:28 crc kubenswrapper[4702]: I1007 12:43:28.587629 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28ce8239b71f08706565ddade636eafe11b67981f199a876e4a97f342ab4fd41" Oct 07 12:43:28 crc kubenswrapper[4702]: I1007 12:43:28.587239 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2rqv8" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.679762 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cb15-account-create-44gnw"] Oct 07 12:43:41 crc kubenswrapper[4702]: E1007 12:43:41.681058 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eaff209-6059-460b-86ed-a577b4abe041" containerName="mariadb-database-create" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.681081 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eaff209-6059-460b-86ed-a577b4abe041" containerName="mariadb-database-create" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.681291 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eaff209-6059-460b-86ed-a577b4abe041" containerName="mariadb-database-create" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.681923 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.684331 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.690690 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cb15-account-create-44gnw"] Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.864709 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrvn4\" (UniqueName: \"kubernetes.io/projected/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8-kube-api-access-xrvn4\") pod \"glance-cb15-account-create-44gnw\" (UID: \"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8\") " pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.965779 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrvn4\" (UniqueName: \"kubernetes.io/projected/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8-kube-api-access-xrvn4\") pod \"glance-cb15-account-create-44gnw\" (UID: \"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8\") " pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:41 crc kubenswrapper[4702]: I1007 12:43:41.993296 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrvn4\" (UniqueName: \"kubernetes.io/projected/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8-kube-api-access-xrvn4\") pod \"glance-cb15-account-create-44gnw\" (UID: \"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8\") " pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:42 crc kubenswrapper[4702]: I1007 12:43:42.005720 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:47 crc kubenswrapper[4702]: I1007 12:43:47.803789 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cb15-account-create-44gnw"] Oct 07 12:43:49 crc kubenswrapper[4702]: I1007 12:43:48.815017 4702 generic.go:334] "Generic (PLEG): container finished" podID="70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8" containerID="d7a73528c3e779b416f4869f51c2d9dd3aadd82148a8b69059676a2837bfe6e3" exitCode=0 Oct 07 12:43:49 crc kubenswrapper[4702]: I1007 12:43:48.815263 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" event={"ID":"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8","Type":"ContainerDied","Data":"d7a73528c3e779b416f4869f51c2d9dd3aadd82148a8b69059676a2837bfe6e3"} Oct 07 12:43:49 crc kubenswrapper[4702]: I1007 12:43:48.816299 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" event={"ID":"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8","Type":"ContainerStarted","Data":"248f022e65aa1e123f6098721751672c50bda1ef2b915ef37dcab49214daa977"} Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.167171 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.307675 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrvn4\" (UniqueName: \"kubernetes.io/projected/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8-kube-api-access-xrvn4\") pod \"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8\" (UID: \"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8\") " Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.318452 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8-kube-api-access-xrvn4" (OuterVolumeSpecName: "kube-api-access-xrvn4") pod "70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8" (UID: "70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8"). InnerVolumeSpecName "kube-api-access-xrvn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.410359 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrvn4\" (UniqueName: \"kubernetes.io/projected/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8-kube-api-access-xrvn4\") on node \"crc\" DevicePath \"\"" Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.836613 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" event={"ID":"70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8","Type":"ContainerDied","Data":"248f022e65aa1e123f6098721751672c50bda1ef2b915ef37dcab49214daa977"} Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.837103 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248f022e65aa1e123f6098721751672c50bda1ef2b915ef37dcab49214daa977" Oct 07 12:43:50 crc kubenswrapper[4702]: I1007 12:43:50.836693 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cb15-account-create-44gnw" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.827766 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-76zhn"] Oct 07 12:43:51 crc kubenswrapper[4702]: E1007 12:43:51.828151 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8" containerName="mariadb-account-create" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.828168 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8" containerName="mariadb-account-create" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.828369 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8" containerName="mariadb-account-create" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.828973 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.837875 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-76zhn"] Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.840424 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.843254 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-zmxx9" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.887722 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-db-sync-config-data\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.887803 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-config-data\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.887837 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj2lz\" (UniqueName: \"kubernetes.io/projected/db7b3179-ec5e-4e64-a189-658a0c34b7d6-kube-api-access-gj2lz\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.990011 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-config-data\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.990074 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj2lz\" (UniqueName: \"kubernetes.io/projected/db7b3179-ec5e-4e64-a189-658a0c34b7d6-kube-api-access-gj2lz\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.990173 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-db-sync-config-data\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.996548 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-db-sync-config-data\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:51 crc kubenswrapper[4702]: I1007 12:43:51.997324 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-config-data\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:52 crc kubenswrapper[4702]: I1007 12:43:52.011376 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj2lz\" (UniqueName: \"kubernetes.io/projected/db7b3179-ec5e-4e64-a189-658a0c34b7d6-kube-api-access-gj2lz\") pod \"glance-db-sync-76zhn\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:52 crc kubenswrapper[4702]: I1007 12:43:52.147557 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:43:52 crc kubenswrapper[4702]: I1007 12:43:52.432777 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-76zhn"] Oct 07 12:43:52 crc kubenswrapper[4702]: W1007 12:43:52.439429 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb7b3179_ec5e_4e64_a189_658a0c34b7d6.slice/crio-27a1600026ae6cf3bcc73f52df42debdeb65cfcbbfa3e515ffa1b082cc3cc856 WatchSource:0}: Error finding container 27a1600026ae6cf3bcc73f52df42debdeb65cfcbbfa3e515ffa1b082cc3cc856: Status 404 returned error can't find the container with id 27a1600026ae6cf3bcc73f52df42debdeb65cfcbbfa3e515ffa1b082cc3cc856 Oct 07 12:43:52 crc kubenswrapper[4702]: I1007 12:43:52.859822 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-76zhn" event={"ID":"db7b3179-ec5e-4e64-a189-658a0c34b7d6","Type":"ContainerStarted","Data":"27a1600026ae6cf3bcc73f52df42debdeb65cfcbbfa3e515ffa1b082cc3cc856"} Oct 07 12:44:06 crc kubenswrapper[4702]: I1007 12:44:06.985138 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-76zhn" event={"ID":"db7b3179-ec5e-4e64-a189-658a0c34b7d6","Type":"ContainerStarted","Data":"d432120e5e9ff79f44484441786c8c50c2520dd944f2a88a69032bbd41379e4f"} Oct 07 12:44:18 crc kubenswrapper[4702]: I1007 12:44:18.106382 4702 generic.go:334] "Generic (PLEG): container finished" podID="db7b3179-ec5e-4e64-a189-658a0c34b7d6" containerID="d432120e5e9ff79f44484441786c8c50c2520dd944f2a88a69032bbd41379e4f" exitCode=0 Oct 07 12:44:18 crc kubenswrapper[4702]: I1007 12:44:18.106475 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-76zhn" event={"ID":"db7b3179-ec5e-4e64-a189-658a0c34b7d6","Type":"ContainerDied","Data":"d432120e5e9ff79f44484441786c8c50c2520dd944f2a88a69032bbd41379e4f"} Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.529252 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.687932 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-db-sync-config-data\") pod \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.688019 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj2lz\" (UniqueName: \"kubernetes.io/projected/db7b3179-ec5e-4e64-a189-658a0c34b7d6-kube-api-access-gj2lz\") pod \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.688129 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-config-data\") pod \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\" (UID: \"db7b3179-ec5e-4e64-a189-658a0c34b7d6\") " Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.715200 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7b3179-ec5e-4e64-a189-658a0c34b7d6-kube-api-access-gj2lz" (OuterVolumeSpecName: "kube-api-access-gj2lz") pod "db7b3179-ec5e-4e64-a189-658a0c34b7d6" (UID: "db7b3179-ec5e-4e64-a189-658a0c34b7d6"). InnerVolumeSpecName "kube-api-access-gj2lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.715695 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "db7b3179-ec5e-4e64-a189-658a0c34b7d6" (UID: "db7b3179-ec5e-4e64-a189-658a0c34b7d6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.795817 4702 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.795865 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj2lz\" (UniqueName: \"kubernetes.io/projected/db7b3179-ec5e-4e64-a189-658a0c34b7d6-kube-api-access-gj2lz\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.797419 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-config-data" (OuterVolumeSpecName: "config-data") pod "db7b3179-ec5e-4e64-a189-658a0c34b7d6" (UID: "db7b3179-ec5e-4e64-a189-658a0c34b7d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:44:19 crc kubenswrapper[4702]: I1007 12:44:19.898019 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7b3179-ec5e-4e64-a189-658a0c34b7d6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.132057 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-76zhn" event={"ID":"db7b3179-ec5e-4e64-a189-658a0c34b7d6","Type":"ContainerDied","Data":"27a1600026ae6cf3bcc73f52df42debdeb65cfcbbfa3e515ffa1b082cc3cc856"} Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.132118 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a1600026ae6cf3bcc73f52df42debdeb65cfcbbfa3e515ffa1b082cc3cc856" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.132255 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-76zhn" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.610727 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:20 crc kubenswrapper[4702]: E1007 12:44:20.611028 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7b3179-ec5e-4e64-a189-658a0c34b7d6" containerName="glance-db-sync" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.611045 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7b3179-ec5e-4e64-a189-658a0c34b7d6" containerName="glance-db-sync" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.611194 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7b3179-ec5e-4e64-a189-658a0c34b7d6" containerName="glance-db-sync" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.611975 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.617383 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.617674 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-zmxx9" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.617814 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.633507 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.634854 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.639423 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.663817 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819017 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-logs\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819091 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kn2z\" (UniqueName: \"kubernetes.io/projected/478e0da4-7ca5-462e-80c3-cdfa1debf860-kube-api-access-5kn2z\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819133 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819167 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-run\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819212 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-config-data\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819268 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-run\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819313 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-nvme\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819356 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-sys\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819384 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819419 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-dev\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819464 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-logs\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819535 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9qg2\" (UniqueName: \"kubernetes.io/projected/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-kube-api-access-c9qg2\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819625 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-lib-modules\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819654 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-httpd-run\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819736 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819774 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-lib-modules\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819807 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819835 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819865 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-dev\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.819976 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820010 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-nvme\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820042 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-httpd-run\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820070 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820098 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-scripts\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820131 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-sys\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820162 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-config-data\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820220 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.820274 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-scripts\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.921853 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-run\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.921936 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-nvme\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.921990 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922010 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-sys\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922031 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-dev\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922062 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-logs\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922069 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-run\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922124 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-nvme\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922124 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-sys\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922089 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9qg2\" (UniqueName: \"kubernetes.io/projected/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-kube-api-access-c9qg2\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922264 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-dev\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922356 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-lib-modules\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922385 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-httpd-run\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922415 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-lib-modules\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922482 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922540 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-lib-modules\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922571 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922599 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-logs\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922601 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922659 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-dev\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922604 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-lib-modules\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922745 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-nvme\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922770 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922804 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-httpd-run\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922827 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922849 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-scripts\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922915 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-sys\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922946 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922934 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-nvme\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922955 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-config-data\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922984 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923034 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-sys\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922977 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922959 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-httpd-run\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.922951 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923185 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923273 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923314 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-scripts\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923350 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kn2z\" (UniqueName: \"kubernetes.io/projected/478e0da4-7ca5-462e-80c3-cdfa1debf860-kube-api-access-5kn2z\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923376 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-logs\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923402 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923432 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-run\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923455 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-config-data\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923554 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-httpd-run\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923697 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.923813 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.925175 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.925177 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-run\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.925450 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-logs\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.925783 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-dev\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.932023 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-config-data\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.933384 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-config-data\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.937383 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-scripts\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.948537 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-scripts\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.948819 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kn2z\" (UniqueName: \"kubernetes.io/projected/478e0da4-7ca5-462e-80c3-cdfa1debf860-kube-api-access-5kn2z\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.954551 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9qg2\" (UniqueName: \"kubernetes.io/projected/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-kube-api-access-c9qg2\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.962104 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.965232 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.966191 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:20 crc kubenswrapper[4702]: I1007 12:44:20.979393 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:21 crc kubenswrapper[4702]: I1007 12:44:21.228753 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:21 crc kubenswrapper[4702]: I1007 12:44:21.258801 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:21 crc kubenswrapper[4702]: I1007 12:44:21.606381 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:21 crc kubenswrapper[4702]: W1007 12:44:21.613429 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod500cc3d8_45ab_4eb4_9e23_6a7d198cd3a5.slice/crio-8fbea0097ba0f8cb0ca3a56ee1119a045537a031f077c5b82654de2cb850805c WatchSource:0}: Error finding container 8fbea0097ba0f8cb0ca3a56ee1119a045537a031f077c5b82654de2cb850805c: Status 404 returned error can't find the container with id 8fbea0097ba0f8cb0ca3a56ee1119a045537a031f077c5b82654de2cb850805c Oct 07 12:44:21 crc kubenswrapper[4702]: I1007 12:44:21.726016 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:21 crc kubenswrapper[4702]: W1007 12:44:21.734355 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478e0da4_7ca5_462e_80c3_cdfa1debf860.slice/crio-1103b83b1f22e8b3ca5f7b94c9cf6ea9be888c65fb6a8cf4292209286e8e9ad4 WatchSource:0}: Error finding container 1103b83b1f22e8b3ca5f7b94c9cf6ea9be888c65fb6a8cf4292209286e8e9ad4: Status 404 returned error can't find the container with id 1103b83b1f22e8b3ca5f7b94c9cf6ea9be888c65fb6a8cf4292209286e8e9ad4 Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.163443 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"478e0da4-7ca5-462e-80c3-cdfa1debf860","Type":"ContainerStarted","Data":"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8"} Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.164218 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"478e0da4-7ca5-462e-80c3-cdfa1debf860","Type":"ContainerStarted","Data":"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00"} Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.164249 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"478e0da4-7ca5-462e-80c3-cdfa1debf860","Type":"ContainerStarted","Data":"1103b83b1f22e8b3ca5f7b94c9cf6ea9be888c65fb6a8cf4292209286e8e9ad4"} Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.165993 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5","Type":"ContainerStarted","Data":"522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e"} Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.166065 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5","Type":"ContainerStarted","Data":"d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa"} Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.166080 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5","Type":"ContainerStarted","Data":"8fbea0097ba0f8cb0ca3a56ee1119a045537a031f077c5b82654de2cb850805c"} Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.192308 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.192286724 podStartE2EDuration="2.192286724s" podCreationTimestamp="2025-10-07 12:44:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:44:22.188664059 +0000 UTC m=+1218.364400205" watchObservedRunningTime="2025-10-07 12:44:22.192286724 +0000 UTC m=+1218.368022840" Oct 07 12:44:22 crc kubenswrapper[4702]: I1007 12:44:22.217478 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.217454693 podStartE2EDuration="2.217454693s" podCreationTimestamp="2025-10-07 12:44:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:44:22.213796407 +0000 UTC m=+1218.389532523" watchObservedRunningTime="2025-10-07 12:44:22.217454693 +0000 UTC m=+1218.393190809" Oct 07 12:44:23 crc kubenswrapper[4702]: I1007 12:44:23.322923 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.199735 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-log" containerID="cri-o://e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00" gracePeriod=30 Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.200117 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-httpd" containerID="cri-o://d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8" gracePeriod=30 Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.758863 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.799956 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-run\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800026 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800078 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-run" (OuterVolumeSpecName: "run") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800097 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-scripts\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800204 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-sys\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800236 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800263 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-config-data\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800288 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-sys" (OuterVolumeSpecName: "sys") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800306 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-httpd-run\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800348 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-lib-modules\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800373 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kn2z\" (UniqueName: \"kubernetes.io/projected/478e0da4-7ca5-462e-80c3-cdfa1debf860-kube-api-access-5kn2z\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800395 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-dev\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800430 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-nvme\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800542 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-var-locks-brick\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800576 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-logs\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800601 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-iscsi\") pod \"478e0da4-7ca5-462e-80c3-cdfa1debf860\" (UID: \"478e0da4-7ca5-462e-80c3-cdfa1debf860\") " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800679 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-dev" (OuterVolumeSpecName: "dev") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800926 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.800965 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801059 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801056 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801105 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801214 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-logs" (OuterVolumeSpecName: "logs") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801273 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-sys\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801289 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801300 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801310 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-dev\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801318 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801326 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.801335 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.808171 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.808588 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.808869 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/478e0da4-7ca5-462e-80c3-cdfa1debf860-kube-api-access-5kn2z" (OuterVolumeSpecName: "kube-api-access-5kn2z") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "kube-api-access-5kn2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.810014 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-scripts" (OuterVolumeSpecName: "scripts") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.853354 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-config-data" (OuterVolumeSpecName: "config-data") pod "478e0da4-7ca5-462e-80c3-cdfa1debf860" (UID: "478e0da4-7ca5-462e-80c3-cdfa1debf860"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902232 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/478e0da4-7ca5-462e-80c3-cdfa1debf860-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902280 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/478e0da4-7ca5-462e-80c3-cdfa1debf860-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902341 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902352 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902366 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902398 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478e0da4-7ca5-462e-80c3-cdfa1debf860-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.902412 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kn2z\" (UniqueName: \"kubernetes.io/projected/478e0da4-7ca5-462e-80c3-cdfa1debf860-kube-api-access-5kn2z\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.918800 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 07 12:44:25 crc kubenswrapper[4702]: I1007 12:44:25.931455 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.003688 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.004230 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219321 4702 generic.go:334] "Generic (PLEG): container finished" podID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerID="d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8" exitCode=0 Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219380 4702 generic.go:334] "Generic (PLEG): container finished" podID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerID="e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00" exitCode=143 Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219417 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"478e0da4-7ca5-462e-80c3-cdfa1debf860","Type":"ContainerDied","Data":"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8"} Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219461 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"478e0da4-7ca5-462e-80c3-cdfa1debf860","Type":"ContainerDied","Data":"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00"} Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219484 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"478e0da4-7ca5-462e-80c3-cdfa1debf860","Type":"ContainerDied","Data":"1103b83b1f22e8b3ca5f7b94c9cf6ea9be888c65fb6a8cf4292209286e8e9ad4"} Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219499 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.219510 4702 scope.go:117] "RemoveContainer" containerID="d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.267804 4702 scope.go:117] "RemoveContainer" containerID="e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.284936 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.296686 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.296711 4702 scope.go:117] "RemoveContainer" containerID="d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8" Oct 07 12:44:26 crc kubenswrapper[4702]: E1007 12:44:26.297445 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8\": container with ID starting with d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8 not found: ID does not exist" containerID="d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.297488 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8"} err="failed to get container status \"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8\": rpc error: code = NotFound desc = could not find container \"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8\": container with ID starting with d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8 not found: ID does not exist" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.297514 4702 scope.go:117] "RemoveContainer" containerID="e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00" Oct 07 12:44:26 crc kubenswrapper[4702]: E1007 12:44:26.297851 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00\": container with ID starting with e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00 not found: ID does not exist" containerID="e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.297877 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00"} err="failed to get container status \"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00\": rpc error: code = NotFound desc = could not find container \"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00\": container with ID starting with e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00 not found: ID does not exist" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.297913 4702 scope.go:117] "RemoveContainer" containerID="d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.298269 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8"} err="failed to get container status \"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8\": rpc error: code = NotFound desc = could not find container \"d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8\": container with ID starting with d52e32afa0b47331038672b25cdc4f9aefcd4fa80a9e995901aa620c6aaa7da8 not found: ID does not exist" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.298292 4702 scope.go:117] "RemoveContainer" containerID="e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.298734 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00"} err="failed to get container status \"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00\": rpc error: code = NotFound desc = could not find container \"e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00\": container with ID starting with e1c72089efd1c076f85841bbf84c00f7bdd98d19726fc384936bcfb7ec744f00 not found: ID does not exist" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.335403 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:26 crc kubenswrapper[4702]: E1007 12:44:26.335801 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-log" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.335825 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-log" Oct 07 12:44:26 crc kubenswrapper[4702]: E1007 12:44:26.335850 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-httpd" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.335858 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-httpd" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.336116 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-log" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.336141 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" containerName="glance-httpd" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.337062 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.347999 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.409816 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-scripts\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410256 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-run\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410302 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-logs\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410351 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-httpd-run\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410392 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-dev\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410419 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410500 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410547 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410576 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-sys\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410612 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-config-data\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410635 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fhfh\" (UniqueName: \"kubernetes.io/projected/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-kube-api-access-2fhfh\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410662 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-nvme\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410687 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.410731 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-lib-modules\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.512857 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.512993 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-sys\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513046 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-config-data\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513079 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fhfh\" (UniqueName: \"kubernetes.io/projected/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-kube-api-access-2fhfh\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513117 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-nvme\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513157 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513217 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-lib-modules\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513257 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-scripts\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513292 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-run\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513337 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-logs\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513398 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-httpd-run\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513416 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513455 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-dev\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513489 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513529 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513697 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513732 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-sys\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.513693 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-lib-modules\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.514884 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-nvme\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.515361 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.515992 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-httpd-run\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.516062 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-run\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.516393 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-logs\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.516455 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-dev\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.516583 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.524058 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-scripts\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.531320 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-config-data\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.538818 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.541742 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fhfh\" (UniqueName: \"kubernetes.io/projected/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-kube-api-access-2fhfh\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.547053 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.653287 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="478e0da4-7ca5-462e-80c3-cdfa1debf860" path="/var/lib/kubelet/pods/478e0da4-7ca5-462e-80c3-cdfa1debf860/volumes" Oct 07 12:44:26 crc kubenswrapper[4702]: I1007 12:44:26.661635 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:27 crc kubenswrapper[4702]: I1007 12:44:27.219649 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:44:27 crc kubenswrapper[4702]: I1007 12:44:27.230993 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"27ebb018-1fe3-4df9-9a90-c08eeeaeb320","Type":"ContainerStarted","Data":"df161965be9d6c52e5947523ca7f3c96476fe34321c10a4e72d606b880368fa1"} Oct 07 12:44:28 crc kubenswrapper[4702]: I1007 12:44:28.244246 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"27ebb018-1fe3-4df9-9a90-c08eeeaeb320","Type":"ContainerStarted","Data":"a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177"} Oct 07 12:44:28 crc kubenswrapper[4702]: I1007 12:44:28.244938 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"27ebb018-1fe3-4df9-9a90-c08eeeaeb320","Type":"ContainerStarted","Data":"13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4"} Oct 07 12:44:28 crc kubenswrapper[4702]: I1007 12:44:28.269209 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.269175323 podStartE2EDuration="2.269175323s" podCreationTimestamp="2025-10-07 12:44:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:44:28.26700037 +0000 UTC m=+1224.442736486" watchObservedRunningTime="2025-10-07 12:44:28.269175323 +0000 UTC m=+1224.444911439" Oct 07 12:44:31 crc kubenswrapper[4702]: I1007 12:44:31.229028 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:31 crc kubenswrapper[4702]: I1007 12:44:31.229348 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:31 crc kubenswrapper[4702]: I1007 12:44:31.264473 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:31 crc kubenswrapper[4702]: I1007 12:44:31.268090 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:31 crc kubenswrapper[4702]: I1007 12:44:31.281341 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:32 crc kubenswrapper[4702]: I1007 12:44:32.277040 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:33 crc kubenswrapper[4702]: I1007 12:44:33.288030 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 12:44:33 crc kubenswrapper[4702]: I1007 12:44:33.355852 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:33 crc kubenswrapper[4702]: I1007 12:44:33.359102 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:36 crc kubenswrapper[4702]: I1007 12:44:36.663218 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:36 crc kubenswrapper[4702]: I1007 12:44:36.663825 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:36 crc kubenswrapper[4702]: I1007 12:44:36.703560 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:36 crc kubenswrapper[4702]: I1007 12:44:36.711785 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:37 crc kubenswrapper[4702]: I1007 12:44:37.321045 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:37 crc kubenswrapper[4702]: I1007 12:44:37.321127 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:39 crc kubenswrapper[4702]: I1007 12:44:39.401754 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:39 crc kubenswrapper[4702]: I1007 12:44:39.402913 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 12:44:39 crc kubenswrapper[4702]: I1007 12:44:39.405064 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:44:39 crc kubenswrapper[4702]: I1007 12:44:39.487167 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:39 crc kubenswrapper[4702]: I1007 12:44:39.487521 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-log" containerID="cri-o://d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa" gracePeriod=30 Oct 07 12:44:39 crc kubenswrapper[4702]: I1007 12:44:39.487616 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-httpd" containerID="cri-o://522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e" gracePeriod=30 Oct 07 12:44:40 crc kubenswrapper[4702]: I1007 12:44:40.352017 4702 generic.go:334] "Generic (PLEG): container finished" podID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerID="d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa" exitCode=143 Oct 07 12:44:40 crc kubenswrapper[4702]: I1007 12:44:40.352143 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5","Type":"ContainerDied","Data":"d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa"} Oct 07 12:44:42 crc kubenswrapper[4702]: E1007 12:44:42.754372 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod500cc3d8_45ab_4eb4_9e23_6a7d198cd3a5.slice/crio-522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e.scope\": RecentStats: unable to find data in memory cache]" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.049330 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.204795 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.204878 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-logs\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.204930 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-nvme\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.204963 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-dev\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.204984 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-config-data\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.204979 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205017 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9qg2\" (UniqueName: \"kubernetes.io/projected/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-kube-api-access-c9qg2\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205041 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-iscsi\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205058 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-lib-modules\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205075 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-sys\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205112 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-scripts\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205127 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205182 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-httpd-run\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205463 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-var-locks-brick\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205487 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-run\") pod \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\" (UID: \"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5\") " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205749 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205038 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-dev" (OuterVolumeSpecName: "dev") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205497 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-logs" (OuterVolumeSpecName: "logs") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205525 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-sys" (OuterVolumeSpecName: "sys") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205791 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-run" (OuterVolumeSpecName: "run") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205976 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.205998 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.206060 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.206340 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.209972 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-scripts" (OuterVolumeSpecName: "scripts") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.210361 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.211178 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-kube-api-access-c9qg2" (OuterVolumeSpecName: "kube-api-access-c9qg2") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "kube-api-access-c9qg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.211785 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.248571 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-config-data" (OuterVolumeSpecName: "config-data") pod "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" (UID: "500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307433 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307472 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307513 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307523 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307534 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-dev\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307560 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307571 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9qg2\" (UniqueName: \"kubernetes.io/projected/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-kube-api-access-c9qg2\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307584 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307592 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307601 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-sys\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307609 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307623 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.307632 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.321373 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.324446 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.379954 4702 generic.go:334] "Generic (PLEG): container finished" podID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerID="522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e" exitCode=0 Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.380000 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5","Type":"ContainerDied","Data":"522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e"} Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.380012 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.380026 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5","Type":"ContainerDied","Data":"8fbea0097ba0f8cb0ca3a56ee1119a045537a031f077c5b82654de2cb850805c"} Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.380046 4702 scope.go:117] "RemoveContainer" containerID="522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.404723 4702 scope.go:117] "RemoveContainer" containerID="d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.409537 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.409575 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.417121 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.424492 4702 scope.go:117] "RemoveContainer" containerID="522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e" Oct 07 12:44:43 crc kubenswrapper[4702]: E1007 12:44:43.425139 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e\": container with ID starting with 522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e not found: ID does not exist" containerID="522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.425198 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e"} err="failed to get container status \"522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e\": rpc error: code = NotFound desc = could not find container \"522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e\": container with ID starting with 522a87fb1b5d30197cb82641703fda7873af369fd64c79c1aaca03b37b74c16e not found: ID does not exist" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.425231 4702 scope.go:117] "RemoveContainer" containerID="d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa" Oct 07 12:44:43 crc kubenswrapper[4702]: E1007 12:44:43.425544 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa\": container with ID starting with d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa not found: ID does not exist" containerID="d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.425583 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa"} err="failed to get container status \"d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa\": rpc error: code = NotFound desc = could not find container \"d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa\": container with ID starting with d9531edcd7c69fab0447d2b0a4342bcc1c45c34ceee83d936550c4345a6f89fa not found: ID does not exist" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.428319 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.452839 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:43 crc kubenswrapper[4702]: E1007 12:44:43.453458 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-log" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.453484 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-log" Oct 07 12:44:43 crc kubenswrapper[4702]: E1007 12:44:43.453510 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-httpd" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.453519 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-httpd" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.453704 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-log" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.453738 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" containerName="glance-httpd" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.454817 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.474912 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612531 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-dev\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612643 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-logs\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612677 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612719 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612735 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-nvme\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612753 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-sys\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612855 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-run\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612929 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-config-data\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612963 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.612994 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-scripts\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.613011 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-httpd-run\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.613026 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b27ks\" (UniqueName: \"kubernetes.io/projected/71e08679-dd5f-48be-9dfb-550aaa330eca-kube-api-access-b27ks\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.613047 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-lib-modules\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.613079 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.714772 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-logs\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715176 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715239 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715259 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-nvme\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-sys\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717053 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-run\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717087 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-config-data\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717131 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717172 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-scripts\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717188 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-httpd-run\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717204 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b27ks\" (UniqueName: \"kubernetes.io/projected/71e08679-dd5f-48be-9dfb-550aaa330eca-kube-api-access-b27ks\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717226 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-lib-modules\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717272 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-dev\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715632 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715521 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-logs\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717593 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-run\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715556 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.717862 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-lib-modules\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.718032 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.718482 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-httpd-run\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715663 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-sys\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.718384 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.718416 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-dev\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.715595 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-nvme\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.723310 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-config-data\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.727561 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-scripts\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.739622 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b27ks\" (UniqueName: \"kubernetes.io/projected/71e08679-dd5f-48be-9dfb-550aaa330eca-kube-api-access-b27ks\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.750060 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.752837 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:43 crc kubenswrapper[4702]: I1007 12:44:43.774692 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:44 crc kubenswrapper[4702]: I1007 12:44:44.201221 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:44:44 crc kubenswrapper[4702]: W1007 12:44:44.213378 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71e08679_dd5f_48be_9dfb_550aaa330eca.slice/crio-409e776f4ce1fabbcb3cc1c5bf5319b78c2f10ae419e794e4398b19fb3a0c3b2 WatchSource:0}: Error finding container 409e776f4ce1fabbcb3cc1c5bf5319b78c2f10ae419e794e4398b19fb3a0c3b2: Status 404 returned error can't find the container with id 409e776f4ce1fabbcb3cc1c5bf5319b78c2f10ae419e794e4398b19fb3a0c3b2 Oct 07 12:44:44 crc kubenswrapper[4702]: I1007 12:44:44.391290 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"71e08679-dd5f-48be-9dfb-550aaa330eca","Type":"ContainerStarted","Data":"409e776f4ce1fabbcb3cc1c5bf5319b78c2f10ae419e794e4398b19fb3a0c3b2"} Oct 07 12:44:44 crc kubenswrapper[4702]: I1007 12:44:44.650542 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5" path="/var/lib/kubelet/pods/500cc3d8-45ab-4eb4-9e23-6a7d198cd3a5/volumes" Oct 07 12:44:45 crc kubenswrapper[4702]: I1007 12:44:45.401402 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"71e08679-dd5f-48be-9dfb-550aaa330eca","Type":"ContainerStarted","Data":"624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520"} Oct 07 12:44:45 crc kubenswrapper[4702]: I1007 12:44:45.403176 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"71e08679-dd5f-48be-9dfb-550aaa330eca","Type":"ContainerStarted","Data":"17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4"} Oct 07 12:44:45 crc kubenswrapper[4702]: I1007 12:44:45.448186 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.448162608 podStartE2EDuration="2.448162608s" podCreationTimestamp="2025-10-07 12:44:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:44:45.440224448 +0000 UTC m=+1241.615960654" watchObservedRunningTime="2025-10-07 12:44:45.448162608 +0000 UTC m=+1241.623898734" Oct 07 12:44:53 crc kubenswrapper[4702]: I1007 12:44:53.775622 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:53 crc kubenswrapper[4702]: I1007 12:44:53.777001 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:53 crc kubenswrapper[4702]: I1007 12:44:53.819765 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:53 crc kubenswrapper[4702]: I1007 12:44:53.834446 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:54 crc kubenswrapper[4702]: I1007 12:44:54.480765 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:54 crc kubenswrapper[4702]: I1007 12:44:54.481134 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:56 crc kubenswrapper[4702]: I1007 12:44:56.509754 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:44:56 crc kubenswrapper[4702]: I1007 12:44:56.510127 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 12:44:56 crc kubenswrapper[4702]: I1007 12:44:56.649960 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.169666 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4"] Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.171717 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.177368 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.177927 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.185333 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4"] Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.339697 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/819fa677-9f32-47c5-8f4d-f543e1582b92-config-volume\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.340295 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/819fa677-9f32-47c5-8f4d-f543e1582b92-secret-volume\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.340485 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hwzh\" (UniqueName: \"kubernetes.io/projected/819fa677-9f32-47c5-8f4d-f543e1582b92-kube-api-access-5hwzh\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.442778 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/819fa677-9f32-47c5-8f4d-f543e1582b92-config-volume\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.443006 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/819fa677-9f32-47c5-8f4d-f543e1582b92-secret-volume\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.443082 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hwzh\" (UniqueName: \"kubernetes.io/projected/819fa677-9f32-47c5-8f4d-f543e1582b92-kube-api-access-5hwzh\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.443973 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/819fa677-9f32-47c5-8f4d-f543e1582b92-config-volume\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.452761 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/819fa677-9f32-47c5-8f4d-f543e1582b92-secret-volume\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.471300 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hwzh\" (UniqueName: \"kubernetes.io/projected/819fa677-9f32-47c5-8f4d-f543e1582b92-kube-api-access-5hwzh\") pod \"collect-profiles-29330685-5ldn4\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:00 crc kubenswrapper[4702]: I1007 12:45:00.548882 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:01 crc kubenswrapper[4702]: I1007 12:45:01.087047 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4"] Oct 07 12:45:01 crc kubenswrapper[4702]: W1007 12:45:01.097757 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod819fa677_9f32_47c5_8f4d_f543e1582b92.slice/crio-63a89d6903a67e68354e4c112262962ea1dd49be0fd49b6239491bcb0484e8cf WatchSource:0}: Error finding container 63a89d6903a67e68354e4c112262962ea1dd49be0fd49b6239491bcb0484e8cf: Status 404 returned error can't find the container with id 63a89d6903a67e68354e4c112262962ea1dd49be0fd49b6239491bcb0484e8cf Oct 07 12:45:01 crc kubenswrapper[4702]: I1007 12:45:01.553486 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" event={"ID":"819fa677-9f32-47c5-8f4d-f543e1582b92","Type":"ContainerStarted","Data":"109448351a347d0273d61c91dfb5bd0267be853f212bd77931bf6a9208b4272f"} Oct 07 12:45:01 crc kubenswrapper[4702]: I1007 12:45:01.553537 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" event={"ID":"819fa677-9f32-47c5-8f4d-f543e1582b92","Type":"ContainerStarted","Data":"63a89d6903a67e68354e4c112262962ea1dd49be0fd49b6239491bcb0484e8cf"} Oct 07 12:45:01 crc kubenswrapper[4702]: I1007 12:45:01.571335 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" podStartSLOduration=1.5713120059999999 podStartE2EDuration="1.571312006s" podCreationTimestamp="2025-10-07 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:45:01.56801726 +0000 UTC m=+1257.743753436" watchObservedRunningTime="2025-10-07 12:45:01.571312006 +0000 UTC m=+1257.747048142" Oct 07 12:45:02 crc kubenswrapper[4702]: I1007 12:45:02.566511 4702 generic.go:334] "Generic (PLEG): container finished" podID="819fa677-9f32-47c5-8f4d-f543e1582b92" containerID="109448351a347d0273d61c91dfb5bd0267be853f212bd77931bf6a9208b4272f" exitCode=0 Oct 07 12:45:02 crc kubenswrapper[4702]: I1007 12:45:02.566654 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" event={"ID":"819fa677-9f32-47c5-8f4d-f543e1582b92","Type":"ContainerDied","Data":"109448351a347d0273d61c91dfb5bd0267be853f212bd77931bf6a9208b4272f"} Oct 07 12:45:03 crc kubenswrapper[4702]: I1007 12:45:03.907434 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.010914 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hwzh\" (UniqueName: \"kubernetes.io/projected/819fa677-9f32-47c5-8f4d-f543e1582b92-kube-api-access-5hwzh\") pod \"819fa677-9f32-47c5-8f4d-f543e1582b92\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.010976 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/819fa677-9f32-47c5-8f4d-f543e1582b92-config-volume\") pod \"819fa677-9f32-47c5-8f4d-f543e1582b92\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.011091 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/819fa677-9f32-47c5-8f4d-f543e1582b92-secret-volume\") pod \"819fa677-9f32-47c5-8f4d-f543e1582b92\" (UID: \"819fa677-9f32-47c5-8f4d-f543e1582b92\") " Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.011842 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819fa677-9f32-47c5-8f4d-f543e1582b92-config-volume" (OuterVolumeSpecName: "config-volume") pod "819fa677-9f32-47c5-8f4d-f543e1582b92" (UID: "819fa677-9f32-47c5-8f4d-f543e1582b92"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.016853 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819fa677-9f32-47c5-8f4d-f543e1582b92-kube-api-access-5hwzh" (OuterVolumeSpecName: "kube-api-access-5hwzh") pod "819fa677-9f32-47c5-8f4d-f543e1582b92" (UID: "819fa677-9f32-47c5-8f4d-f543e1582b92"). InnerVolumeSpecName "kube-api-access-5hwzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.018440 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819fa677-9f32-47c5-8f4d-f543e1582b92-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "819fa677-9f32-47c5-8f4d-f543e1582b92" (UID: "819fa677-9f32-47c5-8f4d-f543e1582b92"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.113324 4702 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/819fa677-9f32-47c5-8f4d-f543e1582b92-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.113646 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hwzh\" (UniqueName: \"kubernetes.io/projected/819fa677-9f32-47c5-8f4d-f543e1582b92-kube-api-access-5hwzh\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.113828 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/819fa677-9f32-47c5-8f4d-f543e1582b92-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.582757 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" event={"ID":"819fa677-9f32-47c5-8f4d-f543e1582b92","Type":"ContainerDied","Data":"63a89d6903a67e68354e4c112262962ea1dd49be0fd49b6239491bcb0484e8cf"} Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.583171 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a89d6903a67e68354e4c112262962ea1dd49be0fd49b6239491bcb0484e8cf" Oct 07 12:45:04 crc kubenswrapper[4702]: I1007 12:45:04.582830 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330685-5ldn4" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.341228 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-76zhn"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.347293 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-76zhn"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.449156 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancecb15-account-delete-2z9s5"] Oct 07 12:45:13 crc kubenswrapper[4702]: E1007 12:45:13.449825 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819fa677-9f32-47c5-8f4d-f543e1582b92" containerName="collect-profiles" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.449842 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="819fa677-9f32-47c5-8f4d-f543e1582b92" containerName="collect-profiles" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.450247 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="819fa677-9f32-47c5-8f4d-f543e1582b92" containerName="collect-profiles" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.451083 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.477598 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.477915 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-log" containerID="cri-o://624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520" gracePeriod=30 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.478100 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-httpd" containerID="cri-o://17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4" gracePeriod=30 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.508714 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.508935 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-log" containerID="cri-o://13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4" gracePeriod=30 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.509265 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-httpd" containerID="cri-o://a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177" gracePeriod=30 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.521117 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancecb15-account-delete-2z9s5"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.549550 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-2rqv8"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.562522 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-2rqv8"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.577839 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancecb15-account-delete-2z9s5"] Oct 07 12:45:13 crc kubenswrapper[4702]: E1007 12:45:13.578524 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-bcd97], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" podUID="dd497a78-d1c9-447b-9348-e8d9859acb68" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.583950 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cb15-account-create-44gnw"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.584534 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcd97\" (UniqueName: \"kubernetes.io/projected/dd497a78-d1c9-447b-9348-e8d9859acb68-kube-api-access-bcd97\") pod \"glancecb15-account-delete-2z9s5\" (UID: \"dd497a78-d1c9-447b-9348-e8d9859acb68\") " pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.587754 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cb15-account-create-44gnw"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.592525 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.592819 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" containerName="openstackclient" containerID="cri-o://8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5" gracePeriod=30 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.661221 4702 generic.go:334] "Generic (PLEG): container finished" podID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerID="13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4" exitCode=143 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.661271 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"27ebb018-1fe3-4df9-9a90-c08eeeaeb320","Type":"ContainerDied","Data":"13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4"} Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.663151 4702 generic.go:334] "Generic (PLEG): container finished" podID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerID="624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520" exitCode=143 Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.663202 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.663655 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"71e08679-dd5f-48be-9dfb-550aaa330eca","Type":"ContainerDied","Data":"624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520"} Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.686116 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcd97\" (UniqueName: \"kubernetes.io/projected/dd497a78-d1c9-447b-9348-e8d9859acb68-kube-api-access-bcd97\") pod \"glancecb15-account-delete-2z9s5\" (UID: \"dd497a78-d1c9-447b-9348-e8d9859acb68\") " pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.706760 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcd97\" (UniqueName: \"kubernetes.io/projected/dd497a78-d1c9-447b-9348-e8d9859acb68-kube-api-access-bcd97\") pod \"glancecb15-account-delete-2z9s5\" (UID: \"dd497a78-d1c9-447b-9348-e8d9859acb68\") " pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.752685 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.888184 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcd97\" (UniqueName: \"kubernetes.io/projected/dd497a78-d1c9-447b-9348-e8d9859acb68-kube-api-access-bcd97\") pod \"dd497a78-d1c9-447b-9348-e8d9859acb68\" (UID: \"dd497a78-d1c9-447b-9348-e8d9859acb68\") " Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.892580 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd497a78-d1c9-447b-9348-e8d9859acb68-kube-api-access-bcd97" (OuterVolumeSpecName: "kube-api-access-bcd97") pod "dd497a78-d1c9-447b-9348-e8d9859acb68" (UID: "dd497a78-d1c9-447b-9348-e8d9859acb68"). InnerVolumeSpecName "kube-api-access-bcd97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:13 crc kubenswrapper[4702]: I1007 12:45:13.989349 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcd97\" (UniqueName: \"kubernetes.io/projected/dd497a78-d1c9-447b-9348-e8d9859acb68-kube-api-access-bcd97\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.022445 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.192992 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgm57\" (UniqueName: \"kubernetes.io/projected/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-kube-api-access-zgm57\") pod \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.193356 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-scripts\") pod \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.193437 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config-secret\") pod \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.193471 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config\") pod \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\" (UID: \"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe\") " Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.194012 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" (UID: "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.198731 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-kube-api-access-zgm57" (OuterVolumeSpecName: "kube-api-access-zgm57") pod "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" (UID: "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe"). InnerVolumeSpecName "kube-api-access-zgm57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.215336 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" (UID: "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.219417 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" (UID: "766bd860-d6e5-4d16-8b5d-3825ec4dd9fe"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.295515 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgm57\" (UniqueName: \"kubernetes.io/projected/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-kube-api-access-zgm57\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.295552 4702 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.295561 4702 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.295570 4702 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.654529 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eaff209-6059-460b-86ed-a577b4abe041" path="/var/lib/kubelet/pods/2eaff209-6059-460b-86ed-a577b4abe041/volumes" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.655217 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8" path="/var/lib/kubelet/pods/70b2bc4c-8295-4aa6-a5d0-e5de99dd7ca8/volumes" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.655667 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7b3179-ec5e-4e64-a189-658a0c34b7d6" path="/var/lib/kubelet/pods/db7b3179-ec5e-4e64-a189-658a0c34b7d6/volumes" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.671331 4702 generic.go:334] "Generic (PLEG): container finished" podID="766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" containerID="8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5" exitCode=143 Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.671375 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe","Type":"ContainerDied","Data":"8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5"} Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.671406 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"766bd860-d6e5-4d16-8b5d-3825ec4dd9fe","Type":"ContainerDied","Data":"234e6350bf57ac21a329a4e9c4bd1fa71d5281518ce6394fe4c37b06f8780723"} Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.671389 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancecb15-account-delete-2z9s5" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.671468 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.671424 4702 scope.go:117] "RemoveContainer" containerID="8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.704637 4702 scope.go:117] "RemoveContainer" containerID="8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5" Oct 07 12:45:14 crc kubenswrapper[4702]: E1007 12:45:14.706167 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5\": container with ID starting with 8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5 not found: ID does not exist" containerID="8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.706217 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5"} err="failed to get container status \"8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5\": rpc error: code = NotFound desc = could not find container \"8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5\": container with ID starting with 8aa3c5e1e4b7ffc282aa08d58b3aae1b27566e53d24f3f77225441e1bec8c4c5 not found: ID does not exist" Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.716652 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancecb15-account-delete-2z9s5"] Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.722606 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancecb15-account-delete-2z9s5"] Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.726999 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:45:14 crc kubenswrapper[4702]: I1007 12:45:14.730528 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:45:16 crc kubenswrapper[4702]: I1007 12:45:16.656522 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" path="/var/lib/kubelet/pods/766bd860-d6e5-4d16-8b5d-3825ec4dd9fe/volumes" Oct 07 12:45:16 crc kubenswrapper[4702]: I1007 12:45:16.657447 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd497a78-d1c9-447b-9348-e8d9859acb68" path="/var/lib/kubelet/pods/dd497a78-d1c9-447b-9348-e8d9859acb68/volumes" Oct 07 12:45:16 crc kubenswrapper[4702]: I1007 12:45:16.679526 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.104:9292/healthcheck\": read tcp 10.217.0.2:54270->10.217.0.104:9292: read: connection reset by peer" Oct 07 12:45:16 crc kubenswrapper[4702]: I1007 12:45:16.679576 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.104:9292/healthcheck\": read tcp 10.217.0.2:54268->10.217.0.104:9292: read: connection reset by peer" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.076716 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.082701 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.243864 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-run\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.243952 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244011 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-httpd-run\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244029 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-config-data\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244030 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-run" (OuterVolumeSpecName: "run") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244051 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-nvme\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244115 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244150 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-config-data\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244207 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-iscsi\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244254 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-scripts\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244287 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-lib-modules\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244323 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fhfh\" (UniqueName: \"kubernetes.io/projected/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-kube-api-access-2fhfh\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244321 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244352 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244365 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-var-locks-brick\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244399 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244420 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-dev\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244454 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-logs\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244478 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-httpd-run\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244498 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244520 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-sys\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244549 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-sys\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244579 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-nvme\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244567 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244618 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-var-locks-brick\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244647 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-dev\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244650 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-sys" (OuterVolumeSpecName: "sys") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244669 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-run\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244690 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-iscsi\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244725 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b27ks\" (UniqueName: \"kubernetes.io/projected/71e08679-dd5f-48be-9dfb-550aaa330eca-kube-api-access-b27ks\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244764 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-scripts\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244803 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-lib-modules\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244824 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-logs\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244854 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"71e08679-dd5f-48be-9dfb-550aaa330eca\" (UID: \"71e08679-dd5f-48be-9dfb-550aaa330eca\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244878 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\" (UID: \"27ebb018-1fe3-4df9-9a90-c08eeeaeb320\") " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.244971 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-dev" (OuterVolumeSpecName: "dev") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245064 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-logs" (OuterVolumeSpecName: "logs") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245089 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245121 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245154 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245160 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-sys" (OuterVolumeSpecName: "sys") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245185 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245629 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245650 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245662 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245674 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245686 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245700 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-dev\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245711 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245725 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245736 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-sys\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245748 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-sys\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245759 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245771 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245782 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245793 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245706 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-logs" (OuterVolumeSpecName: "logs") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245840 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245874 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-dev" (OuterVolumeSpecName: "dev") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.245928 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-run" (OuterVolumeSpecName: "run") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.249384 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-scripts" (OuterVolumeSpecName: "scripts") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.249793 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-kube-api-access-2fhfh" (OuterVolumeSpecName: "kube-api-access-2fhfh") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "kube-api-access-2fhfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.250046 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.250065 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e08679-dd5f-48be-9dfb-550aaa330eca-kube-api-access-b27ks" (OuterVolumeSpecName: "kube-api-access-b27ks") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "kube-api-access-b27ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.250423 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-scripts" (OuterVolumeSpecName: "scripts") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.250503 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.251474 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.251612 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.286250 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-config-data" (OuterVolumeSpecName: "config-data") pod "71e08679-dd5f-48be-9dfb-550aaa330eca" (UID: "71e08679-dd5f-48be-9dfb-550aaa330eca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.293745 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-config-data" (OuterVolumeSpecName: "config-data") pod "27ebb018-1fe3-4df9-9a90-c08eeeaeb320" (UID: "27ebb018-1fe3-4df9-9a90-c08eeeaeb320"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346702 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346733 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/71e08679-dd5f-48be-9dfb-550aaa330eca-dev\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346743 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346751 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b27ks\" (UniqueName: \"kubernetes.io/projected/71e08679-dd5f-48be-9dfb-550aaa330eca-kube-api-access-b27ks\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346760 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346768 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71e08679-dd5f-48be-9dfb-550aaa330eca-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346800 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346813 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346824 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346841 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346850 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e08679-dd5f-48be-9dfb-550aaa330eca-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346858 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346868 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fhfh\" (UniqueName: \"kubernetes.io/projected/27ebb018-1fe3-4df9-9a90-c08eeeaeb320-kube-api-access-2fhfh\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.346880 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.360335 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.361956 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.366192 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.372091 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.447984 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.448031 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.448059 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.448080 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.714680 4702 generic.go:334] "Generic (PLEG): container finished" podID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerID="a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177" exitCode=0 Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.714746 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"27ebb018-1fe3-4df9-9a90-c08eeeaeb320","Type":"ContainerDied","Data":"a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177"} Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.714776 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"27ebb018-1fe3-4df9-9a90-c08eeeaeb320","Type":"ContainerDied","Data":"df161965be9d6c52e5947523ca7f3c96476fe34321c10a4e72d606b880368fa1"} Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.714792 4702 scope.go:117] "RemoveContainer" containerID="a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.714937 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.728711 4702 generic.go:334] "Generic (PLEG): container finished" podID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerID="17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4" exitCode=0 Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.728783 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"71e08679-dd5f-48be-9dfb-550aaa330eca","Type":"ContainerDied","Data":"17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4"} Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.728830 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"71e08679-dd5f-48be-9dfb-550aaa330eca","Type":"ContainerDied","Data":"409e776f4ce1fabbcb3cc1c5bf5319b78c2f10ae419e794e4398b19fb3a0c3b2"} Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.728943 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.777837 4702 scope.go:117] "RemoveContainer" containerID="13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.789490 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.796336 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.802499 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.809277 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.810568 4702 scope.go:117] "RemoveContainer" containerID="a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177" Oct 07 12:45:17 crc kubenswrapper[4702]: E1007 12:45:17.811038 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177\": container with ID starting with a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177 not found: ID does not exist" containerID="a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.811081 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177"} err="failed to get container status \"a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177\": rpc error: code = NotFound desc = could not find container \"a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177\": container with ID starting with a8a4425158576fe1e098f5ffb638d6723e56a95d01b030ee70daf9e0a3e4f177 not found: ID does not exist" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.811111 4702 scope.go:117] "RemoveContainer" containerID="13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4" Oct 07 12:45:17 crc kubenswrapper[4702]: E1007 12:45:17.811518 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4\": container with ID starting with 13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4 not found: ID does not exist" containerID="13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.811548 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4"} err="failed to get container status \"13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4\": rpc error: code = NotFound desc = could not find container \"13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4\": container with ID starting with 13f790666f3856243124c13d25114905de26b282b3f63bcc0f6f218284c963a4 not found: ID does not exist" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.811566 4702 scope.go:117] "RemoveContainer" containerID="17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.848720 4702 scope.go:117] "RemoveContainer" containerID="624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.868611 4702 scope.go:117] "RemoveContainer" containerID="17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4" Oct 07 12:45:17 crc kubenswrapper[4702]: E1007 12:45:17.869177 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4\": container with ID starting with 17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4 not found: ID does not exist" containerID="17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.869220 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4"} err="failed to get container status \"17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4\": rpc error: code = NotFound desc = could not find container \"17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4\": container with ID starting with 17dc7088661b5371ef911eb7212d221a1a2270d23fcde944b442cd12b902d2b4 not found: ID does not exist" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.869248 4702 scope.go:117] "RemoveContainer" containerID="624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520" Oct 07 12:45:17 crc kubenswrapper[4702]: E1007 12:45:17.869573 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520\": container with ID starting with 624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520 not found: ID does not exist" containerID="624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520" Oct 07 12:45:17 crc kubenswrapper[4702]: I1007 12:45:17.869613 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520"} err="failed to get container status \"624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520\": rpc error: code = NotFound desc = could not find container \"624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520\": container with ID starting with 624cbe085c0065bd1271d83ee3a0eeb263b2419d44c5e78f622e8ed7846d5520 not found: ID does not exist" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.613733 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-9jctt"] Oct 07 12:45:18 crc kubenswrapper[4702]: E1007 12:45:18.614332 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-log" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614349 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-log" Oct 07 12:45:18 crc kubenswrapper[4702]: E1007 12:45:18.614366 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-httpd" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614372 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-httpd" Oct 07 12:45:18 crc kubenswrapper[4702]: E1007 12:45:18.614385 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-log" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614392 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-log" Oct 07 12:45:18 crc kubenswrapper[4702]: E1007 12:45:18.614412 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" containerName="openstackclient" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614417 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" containerName="openstackclient" Oct 07 12:45:18 crc kubenswrapper[4702]: E1007 12:45:18.614425 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-httpd" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614431 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-httpd" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614554 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-log" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614568 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="766bd860-d6e5-4d16-8b5d-3825ec4dd9fe" containerName="openstackclient" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614578 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-httpd" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614591 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" containerName="glance-log" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.614601 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" containerName="glance-httpd" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.615111 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.626380 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9jctt"] Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.662426 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27ebb018-1fe3-4df9-9a90-c08eeeaeb320" path="/var/lib/kubelet/pods/27ebb018-1fe3-4df9-9a90-c08eeeaeb320/volumes" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.663549 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71e08679-dd5f-48be-9dfb-550aaa330eca" path="/var/lib/kubelet/pods/71e08679-dd5f-48be-9dfb-550aaa330eca/volumes" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.766512 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bz7k\" (UniqueName: \"kubernetes.io/projected/48a4ce80-94c0-45dd-8bca-9cdbad51387d-kube-api-access-4bz7k\") pod \"glance-db-create-9jctt\" (UID: \"48a4ce80-94c0-45dd-8bca-9cdbad51387d\") " pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.868086 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bz7k\" (UniqueName: \"kubernetes.io/projected/48a4ce80-94c0-45dd-8bca-9cdbad51387d-kube-api-access-4bz7k\") pod \"glance-db-create-9jctt\" (UID: \"48a4ce80-94c0-45dd-8bca-9cdbad51387d\") " pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.907077 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bz7k\" (UniqueName: \"kubernetes.io/projected/48a4ce80-94c0-45dd-8bca-9cdbad51387d-kube-api-access-4bz7k\") pod \"glance-db-create-9jctt\" (UID: \"48a4ce80-94c0-45dd-8bca-9cdbad51387d\") " pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:18 crc kubenswrapper[4702]: I1007 12:45:18.931228 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:19 crc kubenswrapper[4702]: I1007 12:45:19.228296 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9jctt"] Oct 07 12:45:19 crc kubenswrapper[4702]: I1007 12:45:19.745679 4702 generic.go:334] "Generic (PLEG): container finished" podID="48a4ce80-94c0-45dd-8bca-9cdbad51387d" containerID="9436437b6624d68ba66378b440c9ee2e84990c45aff60b742a9d8f2001566ac9" exitCode=0 Oct 07 12:45:19 crc kubenswrapper[4702]: I1007 12:45:19.745744 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9jctt" event={"ID":"48a4ce80-94c0-45dd-8bca-9cdbad51387d","Type":"ContainerDied","Data":"9436437b6624d68ba66378b440c9ee2e84990c45aff60b742a9d8f2001566ac9"} Oct 07 12:45:19 crc kubenswrapper[4702]: I1007 12:45:19.745784 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9jctt" event={"ID":"48a4ce80-94c0-45dd-8bca-9cdbad51387d","Type":"ContainerStarted","Data":"310c9f9f7c1a6de39f4850fd9bbf4be0aa2f4f081671b54aed3705c916816c49"} Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.090395 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.203721 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bz7k\" (UniqueName: \"kubernetes.io/projected/48a4ce80-94c0-45dd-8bca-9cdbad51387d-kube-api-access-4bz7k\") pod \"48a4ce80-94c0-45dd-8bca-9cdbad51387d\" (UID: \"48a4ce80-94c0-45dd-8bca-9cdbad51387d\") " Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.211956 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48a4ce80-94c0-45dd-8bca-9cdbad51387d-kube-api-access-4bz7k" (OuterVolumeSpecName: "kube-api-access-4bz7k") pod "48a4ce80-94c0-45dd-8bca-9cdbad51387d" (UID: "48a4ce80-94c0-45dd-8bca-9cdbad51387d"). InnerVolumeSpecName "kube-api-access-4bz7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.305997 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bz7k\" (UniqueName: \"kubernetes.io/projected/48a4ce80-94c0-45dd-8bca-9cdbad51387d-kube-api-access-4bz7k\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.764310 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9jctt" Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.764295 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9jctt" event={"ID":"48a4ce80-94c0-45dd-8bca-9cdbad51387d","Type":"ContainerDied","Data":"310c9f9f7c1a6de39f4850fd9bbf4be0aa2f4f081671b54aed3705c916816c49"} Oct 07 12:45:21 crc kubenswrapper[4702]: I1007 12:45:21.764917 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="310c9f9f7c1a6de39f4850fd9bbf4be0aa2f4f081671b54aed3705c916816c49" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.668812 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-0211-account-create-mjtsx"] Oct 07 12:45:28 crc kubenswrapper[4702]: E1007 12:45:28.669473 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a4ce80-94c0-45dd-8bca-9cdbad51387d" containerName="mariadb-database-create" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.669489 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a4ce80-94c0-45dd-8bca-9cdbad51387d" containerName="mariadb-database-create" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.669666 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a4ce80-94c0-45dd-8bca-9cdbad51387d" containerName="mariadb-database-create" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.670247 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.680092 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.682326 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0211-account-create-mjtsx"] Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.745775 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fq2n\" (UniqueName: \"kubernetes.io/projected/6a50e790-dc35-484d-beac-9db1b222b62e-kube-api-access-2fq2n\") pod \"glance-0211-account-create-mjtsx\" (UID: \"6a50e790-dc35-484d-beac-9db1b222b62e\") " pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.847215 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fq2n\" (UniqueName: \"kubernetes.io/projected/6a50e790-dc35-484d-beac-9db1b222b62e-kube-api-access-2fq2n\") pod \"glance-0211-account-create-mjtsx\" (UID: \"6a50e790-dc35-484d-beac-9db1b222b62e\") " pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.883143 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fq2n\" (UniqueName: \"kubernetes.io/projected/6a50e790-dc35-484d-beac-9db1b222b62e-kube-api-access-2fq2n\") pod \"glance-0211-account-create-mjtsx\" (UID: \"6a50e790-dc35-484d-beac-9db1b222b62e\") " pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:28 crc kubenswrapper[4702]: I1007 12:45:28.985880 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:29 crc kubenswrapper[4702]: I1007 12:45:29.463723 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0211-account-create-mjtsx"] Oct 07 12:45:29 crc kubenswrapper[4702]: I1007 12:45:29.838282 4702 generic.go:334] "Generic (PLEG): container finished" podID="6a50e790-dc35-484d-beac-9db1b222b62e" containerID="815d9b7862ceaecf7d22ffd5809c84dfb3724c0633100dea2ce54e64903feb50" exitCode=0 Oct 07 12:45:29 crc kubenswrapper[4702]: I1007 12:45:29.838340 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" event={"ID":"6a50e790-dc35-484d-beac-9db1b222b62e","Type":"ContainerDied","Data":"815d9b7862ceaecf7d22ffd5809c84dfb3724c0633100dea2ce54e64903feb50"} Oct 07 12:45:29 crc kubenswrapper[4702]: I1007 12:45:29.838372 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" event={"ID":"6a50e790-dc35-484d-beac-9db1b222b62e","Type":"ContainerStarted","Data":"31b7373623a5d85342690ad9a037472f324a478290114017baeea633a181ce81"} Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.232291 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.385749 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fq2n\" (UniqueName: \"kubernetes.io/projected/6a50e790-dc35-484d-beac-9db1b222b62e-kube-api-access-2fq2n\") pod \"6a50e790-dc35-484d-beac-9db1b222b62e\" (UID: \"6a50e790-dc35-484d-beac-9db1b222b62e\") " Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.393664 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a50e790-dc35-484d-beac-9db1b222b62e-kube-api-access-2fq2n" (OuterVolumeSpecName: "kube-api-access-2fq2n") pod "6a50e790-dc35-484d-beac-9db1b222b62e" (UID: "6a50e790-dc35-484d-beac-9db1b222b62e"). InnerVolumeSpecName "kube-api-access-2fq2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.488361 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fq2n\" (UniqueName: \"kubernetes.io/projected/6a50e790-dc35-484d-beac-9db1b222b62e-kube-api-access-2fq2n\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.856538 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" event={"ID":"6a50e790-dc35-484d-beac-9db1b222b62e","Type":"ContainerDied","Data":"31b7373623a5d85342690ad9a037472f324a478290114017baeea633a181ce81"} Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.856592 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b7373623a5d85342690ad9a037472f324a478290114017baeea633a181ce81" Oct 07 12:45:31 crc kubenswrapper[4702]: I1007 12:45:31.856665 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0211-account-create-mjtsx" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.763579 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-7kpht"] Oct 07 12:45:33 crc kubenswrapper[4702]: E1007 12:45:33.764276 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a50e790-dc35-484d-beac-9db1b222b62e" containerName="mariadb-account-create" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.764298 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a50e790-dc35-484d-beac-9db1b222b62e" containerName="mariadb-account-create" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.764508 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a50e790-dc35-484d-beac-9db1b222b62e" containerName="mariadb-account-create" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.765202 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.768476 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.769047 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.769407 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-tvghn" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.776728 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7kpht"] Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.926630 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf6xj\" (UniqueName: \"kubernetes.io/projected/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-kube-api-access-vf6xj\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.926713 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-config-data\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.926833 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-combined-ca-bundle\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:33 crc kubenswrapper[4702]: I1007 12:45:33.926927 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-db-sync-config-data\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.028503 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-combined-ca-bundle\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.028810 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-db-sync-config-data\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.028965 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf6xj\" (UniqueName: \"kubernetes.io/projected/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-kube-api-access-vf6xj\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.029062 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-config-data\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.037448 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-db-sync-config-data\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.038673 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-config-data\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.038724 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-combined-ca-bundle\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.055219 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf6xj\" (UniqueName: \"kubernetes.io/projected/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-kube-api-access-vf6xj\") pod \"glance-db-sync-7kpht\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.085682 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.405146 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7kpht"] Oct 07 12:45:34 crc kubenswrapper[4702]: W1007 12:45:34.409694 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfc55e53_68f7_4206_bc3d_c57da5d5c22a.slice/crio-19797591a007065fa2386521991d1933b2c4ed69ffb83a2dfad95abfa537b82d WatchSource:0}: Error finding container 19797591a007065fa2386521991d1933b2c4ed69ffb83a2dfad95abfa537b82d: Status 404 returned error can't find the container with id 19797591a007065fa2386521991d1933b2c4ed69ffb83a2dfad95abfa537b82d Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.886431 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7kpht" event={"ID":"cfc55e53-68f7-4206-bc3d-c57da5d5c22a","Type":"ContainerStarted","Data":"ad83c31644dda87324a229c2c270095d7acd9251bea3421771a4291a343a5aaf"} Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.886713 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7kpht" event={"ID":"cfc55e53-68f7-4206-bc3d-c57da5d5c22a","Type":"ContainerStarted","Data":"19797591a007065fa2386521991d1933b2c4ed69ffb83a2dfad95abfa537b82d"} Oct 07 12:45:34 crc kubenswrapper[4702]: I1007 12:45:34.903941 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-7kpht" podStartSLOduration=1.90388706 podStartE2EDuration="1.90388706s" podCreationTimestamp="2025-10-07 12:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:45:34.900043139 +0000 UTC m=+1291.075779295" watchObservedRunningTime="2025-10-07 12:45:34.90388706 +0000 UTC m=+1291.079623186" Oct 07 12:45:37 crc kubenswrapper[4702]: I1007 12:45:37.909342 4702 generic.go:334] "Generic (PLEG): container finished" podID="cfc55e53-68f7-4206-bc3d-c57da5d5c22a" containerID="ad83c31644dda87324a229c2c270095d7acd9251bea3421771a4291a343a5aaf" exitCode=0 Oct 07 12:45:37 crc kubenswrapper[4702]: I1007 12:45:37.909461 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7kpht" event={"ID":"cfc55e53-68f7-4206-bc3d-c57da5d5c22a","Type":"ContainerDied","Data":"ad83c31644dda87324a229c2c270095d7acd9251bea3421771a4291a343a5aaf"} Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.263330 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.263682 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.276930 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.431135 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf6xj\" (UniqueName: \"kubernetes.io/projected/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-kube-api-access-vf6xj\") pod \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.431217 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-config-data\") pod \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.431240 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-db-sync-config-data\") pod \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.431260 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-combined-ca-bundle\") pod \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\" (UID: \"cfc55e53-68f7-4206-bc3d-c57da5d5c22a\") " Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.436500 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-kube-api-access-vf6xj" (OuterVolumeSpecName: "kube-api-access-vf6xj") pod "cfc55e53-68f7-4206-bc3d-c57da5d5c22a" (UID: "cfc55e53-68f7-4206-bc3d-c57da5d5c22a"). InnerVolumeSpecName "kube-api-access-vf6xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.437255 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cfc55e53-68f7-4206-bc3d-c57da5d5c22a" (UID: "cfc55e53-68f7-4206-bc3d-c57da5d5c22a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.451966 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfc55e53-68f7-4206-bc3d-c57da5d5c22a" (UID: "cfc55e53-68f7-4206-bc3d-c57da5d5c22a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.473324 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-config-data" (OuterVolumeSpecName: "config-data") pod "cfc55e53-68f7-4206-bc3d-c57da5d5c22a" (UID: "cfc55e53-68f7-4206-bc3d-c57da5d5c22a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.532613 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.532665 4702 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.532685 4702 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.532703 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf6xj\" (UniqueName: \"kubernetes.io/projected/cfc55e53-68f7-4206-bc3d-c57da5d5c22a-kube-api-access-vf6xj\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.931933 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7kpht" event={"ID":"cfc55e53-68f7-4206-bc3d-c57da5d5c22a","Type":"ContainerDied","Data":"19797591a007065fa2386521991d1933b2c4ed69ffb83a2dfad95abfa537b82d"} Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.931973 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7kpht" Oct 07 12:45:39 crc kubenswrapper[4702]: I1007 12:45:39.931987 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19797591a007065fa2386521991d1933b2c4ed69ffb83a2dfad95abfa537b82d" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.319165 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:40 crc kubenswrapper[4702]: E1007 12:45:40.319625 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc55e53-68f7-4206-bc3d-c57da5d5c22a" containerName="glance-db-sync" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.319649 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc55e53-68f7-4206-bc3d-c57da5d5c22a" containerName="glance-db-sync" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.319930 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc55e53-68f7-4206-bc3d-c57da5d5c22a" containerName="glance-db-sync" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.321082 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.322750 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.322833 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-tvghn" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.323833 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.325961 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.326348 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.326403 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.326468 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.342867 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:40 crc kubenswrapper[4702]: E1007 12:45:40.343565 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-5xkjp logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-5xkjp logs public-tls-certs scripts]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.448630 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-logs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.448683 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.448867 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xkjp\" (UniqueName: \"kubernetes.io/projected/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-kube-api-access-5xkjp\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.448965 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.449037 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-config-data\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.449135 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.449243 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.449320 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-httpd-run\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.449365 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-scripts\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550288 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550351 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xkjp\" (UniqueName: \"kubernetes.io/projected/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-kube-api-access-5xkjp\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550378 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550406 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-config-data\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550435 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550467 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550495 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-httpd-run\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550510 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-scripts\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.550780 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.551020 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-httpd-run\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.551145 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-logs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.551474 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-logs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.559502 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.559584 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.562610 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-scripts\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.563068 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-config-data\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.575491 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.586857 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.586938 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xkjp\" (UniqueName: \"kubernetes.io/projected/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-kube-api-access-5xkjp\") pod \"glance-default-single-0\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.937987 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:40 crc kubenswrapper[4702]: I1007 12:45:40.946311 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057308 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-config-data\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057361 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057389 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-public-tls-certs\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057411 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xkjp\" (UniqueName: \"kubernetes.io/projected/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-kube-api-access-5xkjp\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057450 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-scripts\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057510 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-combined-ca-bundle\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057555 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-logs\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057577 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-httpd-run\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.057620 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-internal-tls-certs\") pod \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\" (UID: \"b5a5dd5c-bdf5-4639-baa7-63e4edfb905d\") " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.058108 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.058124 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-logs" (OuterVolumeSpecName: "logs") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.060916 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.062120 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-kube-api-access-5xkjp" (OuterVolumeSpecName: "kube-api-access-5xkjp") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "kube-api-access-5xkjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.062234 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.062486 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.065008 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.066169 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-config-data" (OuterVolumeSpecName: "config-data") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.066387 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-scripts" (OuterVolumeSpecName: "scripts") pod "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" (UID: "b5a5dd5c-bdf5-4639-baa7-63e4edfb905d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158674 4702 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158707 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158718 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158726 4702 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158734 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158764 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158773 4702 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158781 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xkjp\" (UniqueName: \"kubernetes.io/projected/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-kube-api-access-5xkjp\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.158790 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.170852 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.260513 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.944769 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:41 crc kubenswrapper[4702]: I1007 12:45:41.995869 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:42 crc kubenswrapper[4702]: I1007 12:45:42.001453 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:42 crc kubenswrapper[4702]: I1007 12:45:42.655710 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5a5dd5c-bdf5-4639-baa7-63e4edfb905d" path="/var/lib/kubelet/pods/b5a5dd5c-bdf5-4639-baa7-63e4edfb905d/volumes" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.921149 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.923305 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.926319 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.927424 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.928287 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-tvghn" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.928428 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.928485 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.928996 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 07 12:45:46 crc kubenswrapper[4702]: I1007 12:45:46.944380 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.099593 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-scripts\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.099956 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.099989 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-config-data\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.100070 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.100166 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.100224 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.100246 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv97k\" (UniqueName: \"kubernetes.io/projected/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-kube-api-access-hv97k\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.100328 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-httpd-run\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.100367 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-logs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202117 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-scripts\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202231 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202294 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-config-data\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202341 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202388 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202436 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202473 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv97k\" (UniqueName: \"kubernetes.io/projected/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-kube-api-access-hv97k\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202556 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-httpd-run\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.202630 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-logs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.203669 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.203667 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-logs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.205149 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-httpd-run\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.209981 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-scripts\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.210736 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.210867 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.211080 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-config-data\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.211531 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.228618 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv97k\" (UniqueName: \"kubernetes.io/projected/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-kube-api-access-hv97k\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.233479 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.240785 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:47 crc kubenswrapper[4702]: I1007 12:45:47.712162 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:45:48 crc kubenswrapper[4702]: I1007 12:45:48.013747 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f","Type":"ContainerStarted","Data":"806fd6e1ee514d0b0909b3823c3c1f4baedc770ec6431137336a38d995c4163a"} Oct 07 12:45:49 crc kubenswrapper[4702]: I1007 12:45:49.023145 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f","Type":"ContainerStarted","Data":"f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66"} Oct 07 12:45:49 crc kubenswrapper[4702]: I1007 12:45:49.023430 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f","Type":"ContainerStarted","Data":"45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189"} Oct 07 12:45:49 crc kubenswrapper[4702]: I1007 12:45:49.048509 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.048487391 podStartE2EDuration="3.048487391s" podCreationTimestamp="2025-10-07 12:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:45:49.044891477 +0000 UTC m=+1305.220627613" watchObservedRunningTime="2025-10-07 12:45:49.048487391 +0000 UTC m=+1305.224223507" Oct 07 12:45:57 crc kubenswrapper[4702]: I1007 12:45:57.242093 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:57 crc kubenswrapper[4702]: I1007 12:45:57.243580 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:57 crc kubenswrapper[4702]: I1007 12:45:57.284080 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:57 crc kubenswrapper[4702]: I1007 12:45:57.300475 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:58 crc kubenswrapper[4702]: I1007 12:45:58.124050 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:58 crc kubenswrapper[4702]: I1007 12:45:58.124426 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:45:59 crc kubenswrapper[4702]: I1007 12:45:59.970245 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:46:00 crc kubenswrapper[4702]: I1007 12:46:00.010599 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.333824 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7kpht"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.349277 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7kpht"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.416753 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.448294 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance0211-account-delete-brnwk"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.449362 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.457958 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance0211-account-delete-brnwk"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.502503 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-9jctt"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.514847 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-9jctt"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.526822 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-0211-account-create-mjtsx"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.532302 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance0211-account-delete-brnwk"] Oct 07 12:46:01 crc kubenswrapper[4702]: E1007 12:46:01.532535 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-hmpc2], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance0211-account-delete-brnwk" podUID="79734547-d2c0-4839-8033-10105dbfac12" Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.538606 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-0211-account-create-mjtsx"] Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.567927 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmpc2\" (UniqueName: \"kubernetes.io/projected/79734547-d2c0-4839-8033-10105dbfac12-kube-api-access-hmpc2\") pod \"glance0211-account-delete-brnwk\" (UID: \"79734547-d2c0-4839-8033-10105dbfac12\") " pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.669907 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmpc2\" (UniqueName: \"kubernetes.io/projected/79734547-d2c0-4839-8033-10105dbfac12-kube-api-access-hmpc2\") pod \"glance0211-account-delete-brnwk\" (UID: \"79734547-d2c0-4839-8033-10105dbfac12\") " pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:01 crc kubenswrapper[4702]: I1007 12:46:01.688763 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmpc2\" (UniqueName: \"kubernetes.io/projected/79734547-d2c0-4839-8033-10105dbfac12-kube-api-access-hmpc2\") pod \"glance0211-account-delete-brnwk\" (UID: \"79734547-d2c0-4839-8033-10105dbfac12\") " pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.162864 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.163166 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-log" containerID="cri-o://45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189" gracePeriod=30 Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.163332 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-httpd" containerID="cri-o://f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66" gracePeriod=30 Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.182583 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.278580 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmpc2\" (UniqueName: \"kubernetes.io/projected/79734547-d2c0-4839-8033-10105dbfac12-kube-api-access-hmpc2\") pod \"79734547-d2c0-4839-8033-10105dbfac12\" (UID: \"79734547-d2c0-4839-8033-10105dbfac12\") " Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.282422 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79734547-d2c0-4839-8033-10105dbfac12-kube-api-access-hmpc2" (OuterVolumeSpecName: "kube-api-access-hmpc2") pod "79734547-d2c0-4839-8033-10105dbfac12" (UID: "79734547-d2c0-4839-8033-10105dbfac12"). InnerVolumeSpecName "kube-api-access-hmpc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.380814 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmpc2\" (UniqueName: \"kubernetes.io/projected/79734547-d2c0-4839-8033-10105dbfac12-kube-api-access-hmpc2\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.656701 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48a4ce80-94c0-45dd-8bca-9cdbad51387d" path="/var/lib/kubelet/pods/48a4ce80-94c0-45dd-8bca-9cdbad51387d/volumes" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.658263 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a50e790-dc35-484d-beac-9db1b222b62e" path="/var/lib/kubelet/pods/6a50e790-dc35-484d-beac-9db1b222b62e/volumes" Oct 07 12:46:02 crc kubenswrapper[4702]: I1007 12:46:02.659465 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc55e53-68f7-4206-bc3d-c57da5d5c22a" path="/var/lib/kubelet/pods/cfc55e53-68f7-4206-bc3d-c57da5d5c22a/volumes" Oct 07 12:46:03 crc kubenswrapper[4702]: I1007 12:46:03.173145 4702 generic.go:334] "Generic (PLEG): container finished" podID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerID="45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189" exitCode=143 Oct 07 12:46:03 crc kubenswrapper[4702]: I1007 12:46:03.173256 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0211-account-delete-brnwk" Oct 07 12:46:03 crc kubenswrapper[4702]: I1007 12:46:03.173248 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f","Type":"ContainerDied","Data":"45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189"} Oct 07 12:46:03 crc kubenswrapper[4702]: I1007 12:46:03.207359 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance0211-account-delete-brnwk"] Oct 07 12:46:03 crc kubenswrapper[4702]: I1007 12:46:03.221094 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance0211-account-delete-brnwk"] Oct 07 12:46:04 crc kubenswrapper[4702]: I1007 12:46:04.659981 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79734547-d2c0-4839-8033-10105dbfac12" path="/var/lib/kubelet/pods/79734547-d2c0-4839-8033-10105dbfac12/volumes" Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.828934 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.939772 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-scripts\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.940919 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-combined-ca-bundle\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941036 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-config-data\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941084 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-public-tls-certs\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941140 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-internal-tls-certs\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941214 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-logs\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941257 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv97k\" (UniqueName: \"kubernetes.io/projected/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-kube-api-access-hv97k\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941294 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-httpd-run\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.941366 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\" (UID: \"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f\") " Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.947881 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-logs" (OuterVolumeSpecName: "logs") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.963473 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.967111 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-scripts" (OuterVolumeSpecName: "scripts") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.973045 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:46:05 crc kubenswrapper[4702]: I1007 12:46:05.975097 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-kube-api-access-hv97k" (OuterVolumeSpecName: "kube-api-access-hv97k") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "kube-api-access-hv97k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.014061 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.043317 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.043346 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.043357 4702 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.043367 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.043376 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv97k\" (UniqueName: \"kubernetes.io/projected/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-kube-api-access-hv97k\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.043385 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.050710 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-config-data" (OuterVolumeSpecName: "config-data") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.053224 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.055244 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" (UID: "fe3e29e1-fa64-4f36-98fd-4acb4c57b63f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.057171 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.144837 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.144883 4702 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.144921 4702 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.144937 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.213455 4702 generic.go:334] "Generic (PLEG): container finished" podID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerID="f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66" exitCode=0 Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.213506 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f","Type":"ContainerDied","Data":"f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66"} Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.213537 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"fe3e29e1-fa64-4f36-98fd-4acb4c57b63f","Type":"ContainerDied","Data":"806fd6e1ee514d0b0909b3823c3c1f4baedc770ec6431137336a38d995c4163a"} Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.213556 4702 scope.go:117] "RemoveContainer" containerID="f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.213677 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.256100 4702 scope.go:117] "RemoveContainer" containerID="45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.261029 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.267196 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.282300 4702 scope.go:117] "RemoveContainer" containerID="f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66" Oct 07 12:46:06 crc kubenswrapper[4702]: E1007 12:46:06.283038 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66\": container with ID starting with f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66 not found: ID does not exist" containerID="f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.283070 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66"} err="failed to get container status \"f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66\": rpc error: code = NotFound desc = could not find container \"f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66\": container with ID starting with f221ecae3023c0a7435bcaa54df08a0aec73b349a11589e345afac9e0bde1a66 not found: ID does not exist" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.283089 4702 scope.go:117] "RemoveContainer" containerID="45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189" Oct 07 12:46:06 crc kubenswrapper[4702]: E1007 12:46:06.283651 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189\": container with ID starting with 45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189 not found: ID does not exist" containerID="45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.283699 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189"} err="failed to get container status \"45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189\": rpc error: code = NotFound desc = could not find container \"45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189\": container with ID starting with 45c83cde38112a1eb6ecaac51820b7b24ddd8434bdda428caaaa3bb61001f189 not found: ID does not exist" Oct 07 12:46:06 crc kubenswrapper[4702]: I1007 12:46:06.657185 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" path="/var/lib/kubelet/pods/fe3e29e1-fa64-4f36-98fd-4acb4c57b63f/volumes" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.166555 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-6kvrs"] Oct 07 12:46:07 crc kubenswrapper[4702]: E1007 12:46:07.167285 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-log" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.167301 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-log" Oct 07 12:46:07 crc kubenswrapper[4702]: E1007 12:46:07.167317 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-httpd" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.167326 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-httpd" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.167512 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-httpd" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.167536 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3e29e1-fa64-4f36-98fd-4acb4c57b63f" containerName="glance-log" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.168090 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.189495 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-6kvrs"] Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.264416 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnpzg\" (UniqueName: \"kubernetes.io/projected/598ee324-eed2-48a2-b64d-2e439a2002af-kube-api-access-qnpzg\") pod \"glance-db-create-6kvrs\" (UID: \"598ee324-eed2-48a2-b64d-2e439a2002af\") " pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.365265 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnpzg\" (UniqueName: \"kubernetes.io/projected/598ee324-eed2-48a2-b64d-2e439a2002af-kube-api-access-qnpzg\") pod \"glance-db-create-6kvrs\" (UID: \"598ee324-eed2-48a2-b64d-2e439a2002af\") " pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.390053 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnpzg\" (UniqueName: \"kubernetes.io/projected/598ee324-eed2-48a2-b64d-2e439a2002af-kube-api-access-qnpzg\") pod \"glance-db-create-6kvrs\" (UID: \"598ee324-eed2-48a2-b64d-2e439a2002af\") " pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:07 crc kubenswrapper[4702]: I1007 12:46:07.496134 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:08 crc kubenswrapper[4702]: I1007 12:46:08.014843 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-6kvrs"] Oct 07 12:46:08 crc kubenswrapper[4702]: I1007 12:46:08.248797 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6kvrs" event={"ID":"598ee324-eed2-48a2-b64d-2e439a2002af","Type":"ContainerStarted","Data":"9997cce2b6960ff81d9e8af96d39546a9d7c65142dea8b839e47443813e725c0"} Oct 07 12:46:08 crc kubenswrapper[4702]: I1007 12:46:08.248852 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6kvrs" event={"ID":"598ee324-eed2-48a2-b64d-2e439a2002af","Type":"ContainerStarted","Data":"4515a0326c39517e8c99a16ea8f2a43eadb3e3a4c83c74f493c3d1ac031173c3"} Oct 07 12:46:08 crc kubenswrapper[4702]: I1007 12:46:08.267961 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-6kvrs" podStartSLOduration=1.267945502 podStartE2EDuration="1.267945502s" podCreationTimestamp="2025-10-07 12:46:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:46:08.266214332 +0000 UTC m=+1324.441950448" watchObservedRunningTime="2025-10-07 12:46:08.267945502 +0000 UTC m=+1324.443681618" Oct 07 12:46:09 crc kubenswrapper[4702]: I1007 12:46:09.260237 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6kvrs" event={"ID":"598ee324-eed2-48a2-b64d-2e439a2002af","Type":"ContainerDied","Data":"9997cce2b6960ff81d9e8af96d39546a9d7c65142dea8b839e47443813e725c0"} Oct 07 12:46:09 crc kubenswrapper[4702]: I1007 12:46:09.260598 4702 generic.go:334] "Generic (PLEG): container finished" podID="598ee324-eed2-48a2-b64d-2e439a2002af" containerID="9997cce2b6960ff81d9e8af96d39546a9d7c65142dea8b839e47443813e725c0" exitCode=0 Oct 07 12:46:09 crc kubenswrapper[4702]: I1007 12:46:09.262280 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:46:09 crc kubenswrapper[4702]: I1007 12:46:09.262443 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:46:10 crc kubenswrapper[4702]: I1007 12:46:10.667532 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:10 crc kubenswrapper[4702]: I1007 12:46:10.819760 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnpzg\" (UniqueName: \"kubernetes.io/projected/598ee324-eed2-48a2-b64d-2e439a2002af-kube-api-access-qnpzg\") pod \"598ee324-eed2-48a2-b64d-2e439a2002af\" (UID: \"598ee324-eed2-48a2-b64d-2e439a2002af\") " Oct 07 12:46:10 crc kubenswrapper[4702]: I1007 12:46:10.828181 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/598ee324-eed2-48a2-b64d-2e439a2002af-kube-api-access-qnpzg" (OuterVolumeSpecName: "kube-api-access-qnpzg") pod "598ee324-eed2-48a2-b64d-2e439a2002af" (UID: "598ee324-eed2-48a2-b64d-2e439a2002af"). InnerVolumeSpecName "kube-api-access-qnpzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:46:10 crc kubenswrapper[4702]: I1007 12:46:10.922145 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnpzg\" (UniqueName: \"kubernetes.io/projected/598ee324-eed2-48a2-b64d-2e439a2002af-kube-api-access-qnpzg\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:11 crc kubenswrapper[4702]: I1007 12:46:11.285407 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-6kvrs" event={"ID":"598ee324-eed2-48a2-b64d-2e439a2002af","Type":"ContainerDied","Data":"4515a0326c39517e8c99a16ea8f2a43eadb3e3a4c83c74f493c3d1ac031173c3"} Oct 07 12:46:11 crc kubenswrapper[4702]: I1007 12:46:11.285468 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-6kvrs" Oct 07 12:46:11 crc kubenswrapper[4702]: I1007 12:46:11.285474 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4515a0326c39517e8c99a16ea8f2a43eadb3e3a4c83c74f493c3d1ac031173c3" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.199932 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-7db2-account-create-mkb7h"] Oct 07 12:46:17 crc kubenswrapper[4702]: E1007 12:46:17.201112 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598ee324-eed2-48a2-b64d-2e439a2002af" containerName="mariadb-database-create" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.201129 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="598ee324-eed2-48a2-b64d-2e439a2002af" containerName="mariadb-database-create" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.201448 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="598ee324-eed2-48a2-b64d-2e439a2002af" containerName="mariadb-database-create" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.203127 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.216400 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.219153 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-7db2-account-create-mkb7h"] Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.228913 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfgzn\" (UniqueName: \"kubernetes.io/projected/1c75a1f9-755f-4b5e-9e45-88fb60ce3596-kube-api-access-pfgzn\") pod \"glance-7db2-account-create-mkb7h\" (UID: \"1c75a1f9-755f-4b5e-9e45-88fb60ce3596\") " pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.330948 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfgzn\" (UniqueName: \"kubernetes.io/projected/1c75a1f9-755f-4b5e-9e45-88fb60ce3596-kube-api-access-pfgzn\") pod \"glance-7db2-account-create-mkb7h\" (UID: \"1c75a1f9-755f-4b5e-9e45-88fb60ce3596\") " pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.360552 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfgzn\" (UniqueName: \"kubernetes.io/projected/1c75a1f9-755f-4b5e-9e45-88fb60ce3596-kube-api-access-pfgzn\") pod \"glance-7db2-account-create-mkb7h\" (UID: \"1c75a1f9-755f-4b5e-9e45-88fb60ce3596\") " pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.545053 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:17 crc kubenswrapper[4702]: I1007 12:46:17.989210 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-7db2-account-create-mkb7h"] Oct 07 12:46:18 crc kubenswrapper[4702]: I1007 12:46:18.362748 4702 generic.go:334] "Generic (PLEG): container finished" podID="1c75a1f9-755f-4b5e-9e45-88fb60ce3596" containerID="61a1d71a3dfc9f1d2b96dd869abc1fc2a28836d26b627c02cf4eecdc3cc6c263" exitCode=0 Oct 07 12:46:18 crc kubenswrapper[4702]: I1007 12:46:18.362791 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" event={"ID":"1c75a1f9-755f-4b5e-9e45-88fb60ce3596","Type":"ContainerDied","Data":"61a1d71a3dfc9f1d2b96dd869abc1fc2a28836d26b627c02cf4eecdc3cc6c263"} Oct 07 12:46:18 crc kubenswrapper[4702]: I1007 12:46:18.362821 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" event={"ID":"1c75a1f9-755f-4b5e-9e45-88fb60ce3596","Type":"ContainerStarted","Data":"90160c621f57d70a2cdc8953c99c11dc2785730d1330826ba4072f8d25645fea"} Oct 07 12:46:19 crc kubenswrapper[4702]: I1007 12:46:19.804957 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:19 crc kubenswrapper[4702]: I1007 12:46:19.965077 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfgzn\" (UniqueName: \"kubernetes.io/projected/1c75a1f9-755f-4b5e-9e45-88fb60ce3596-kube-api-access-pfgzn\") pod \"1c75a1f9-755f-4b5e-9e45-88fb60ce3596\" (UID: \"1c75a1f9-755f-4b5e-9e45-88fb60ce3596\") " Oct 07 12:46:19 crc kubenswrapper[4702]: I1007 12:46:19.972122 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c75a1f9-755f-4b5e-9e45-88fb60ce3596-kube-api-access-pfgzn" (OuterVolumeSpecName: "kube-api-access-pfgzn") pod "1c75a1f9-755f-4b5e-9e45-88fb60ce3596" (UID: "1c75a1f9-755f-4b5e-9e45-88fb60ce3596"). InnerVolumeSpecName "kube-api-access-pfgzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:46:20 crc kubenswrapper[4702]: I1007 12:46:20.067536 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfgzn\" (UniqueName: \"kubernetes.io/projected/1c75a1f9-755f-4b5e-9e45-88fb60ce3596-kube-api-access-pfgzn\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:20 crc kubenswrapper[4702]: I1007 12:46:20.384239 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" event={"ID":"1c75a1f9-755f-4b5e-9e45-88fb60ce3596","Type":"ContainerDied","Data":"90160c621f57d70a2cdc8953c99c11dc2785730d1330826ba4072f8d25645fea"} Oct 07 12:46:20 crc kubenswrapper[4702]: I1007 12:46:20.384303 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90160c621f57d70a2cdc8953c99c11dc2785730d1330826ba4072f8d25645fea" Oct 07 12:46:20 crc kubenswrapper[4702]: I1007 12:46:20.384368 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-7db2-account-create-mkb7h" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.358431 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-98k5l"] Oct 07 12:46:22 crc kubenswrapper[4702]: E1007 12:46:22.359131 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75a1f9-755f-4b5e-9e45-88fb60ce3596" containerName="mariadb-account-create" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.359153 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75a1f9-755f-4b5e-9e45-88fb60ce3596" containerName="mariadb-account-create" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.359411 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c75a1f9-755f-4b5e-9e45-88fb60ce3596" containerName="mariadb-account-create" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.360153 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.364361 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.365308 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-vv9tn" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.370634 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-98k5l"] Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.410539 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-db-sync-config-data\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.410596 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvfnp\" (UniqueName: \"kubernetes.io/projected/9a483967-d369-48b0-8e8b-0b38b010e771-kube-api-access-tvfnp\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.410660 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-config-data\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.511423 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvfnp\" (UniqueName: \"kubernetes.io/projected/9a483967-d369-48b0-8e8b-0b38b010e771-kube-api-access-tvfnp\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.511505 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-config-data\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.511576 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-db-sync-config-data\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.517046 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-db-sync-config-data\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.531132 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-config-data\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.545089 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvfnp\" (UniqueName: \"kubernetes.io/projected/9a483967-d369-48b0-8e8b-0b38b010e771-kube-api-access-tvfnp\") pod \"glance-db-sync-98k5l\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:22 crc kubenswrapper[4702]: I1007 12:46:22.684275 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:23 crc kubenswrapper[4702]: I1007 12:46:23.142809 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-98k5l"] Oct 07 12:46:23 crc kubenswrapper[4702]: W1007 12:46:23.144893 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a483967_d369_48b0_8e8b_0b38b010e771.slice/crio-cf33e57073ccb91521f88bec1fd582a5c52bf47a210f66345a3caf845cd3e9c9 WatchSource:0}: Error finding container cf33e57073ccb91521f88bec1fd582a5c52bf47a210f66345a3caf845cd3e9c9: Status 404 returned error can't find the container with id cf33e57073ccb91521f88bec1fd582a5c52bf47a210f66345a3caf845cd3e9c9 Oct 07 12:46:23 crc kubenswrapper[4702]: I1007 12:46:23.409583 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-98k5l" event={"ID":"9a483967-d369-48b0-8e8b-0b38b010e771","Type":"ContainerStarted","Data":"cf33e57073ccb91521f88bec1fd582a5c52bf47a210f66345a3caf845cd3e9c9"} Oct 07 12:46:24 crc kubenswrapper[4702]: I1007 12:46:24.436493 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-98k5l" event={"ID":"9a483967-d369-48b0-8e8b-0b38b010e771","Type":"ContainerStarted","Data":"b2afe95112d7ef6a6054011ec90bf2d3e9fab48538ce3e2d1ee5ee5792ac256c"} Oct 07 12:46:24 crc kubenswrapper[4702]: I1007 12:46:24.466688 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-98k5l" podStartSLOduration=2.466659831 podStartE2EDuration="2.466659831s" podCreationTimestamp="2025-10-07 12:46:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:46:24.457002153 +0000 UTC m=+1340.632738269" watchObservedRunningTime="2025-10-07 12:46:24.466659831 +0000 UTC m=+1340.642395987" Oct 07 12:46:27 crc kubenswrapper[4702]: I1007 12:46:27.475464 4702 generic.go:334] "Generic (PLEG): container finished" podID="9a483967-d369-48b0-8e8b-0b38b010e771" containerID="b2afe95112d7ef6a6054011ec90bf2d3e9fab48538ce3e2d1ee5ee5792ac256c" exitCode=0 Oct 07 12:46:27 crc kubenswrapper[4702]: I1007 12:46:27.475584 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-98k5l" event={"ID":"9a483967-d369-48b0-8e8b-0b38b010e771","Type":"ContainerDied","Data":"b2afe95112d7ef6a6054011ec90bf2d3e9fab48538ce3e2d1ee5ee5792ac256c"} Oct 07 12:46:28 crc kubenswrapper[4702]: I1007 12:46:28.831128 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.013352 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-db-sync-config-data\") pod \"9a483967-d369-48b0-8e8b-0b38b010e771\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.013427 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-config-data\") pod \"9a483967-d369-48b0-8e8b-0b38b010e771\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.013557 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvfnp\" (UniqueName: \"kubernetes.io/projected/9a483967-d369-48b0-8e8b-0b38b010e771-kube-api-access-tvfnp\") pod \"9a483967-d369-48b0-8e8b-0b38b010e771\" (UID: \"9a483967-d369-48b0-8e8b-0b38b010e771\") " Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.020213 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9a483967-d369-48b0-8e8b-0b38b010e771" (UID: "9a483967-d369-48b0-8e8b-0b38b010e771"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.033692 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a483967-d369-48b0-8e8b-0b38b010e771-kube-api-access-tvfnp" (OuterVolumeSpecName: "kube-api-access-tvfnp") pod "9a483967-d369-48b0-8e8b-0b38b010e771" (UID: "9a483967-d369-48b0-8e8b-0b38b010e771"). InnerVolumeSpecName "kube-api-access-tvfnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.055799 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-config-data" (OuterVolumeSpecName: "config-data") pod "9a483967-d369-48b0-8e8b-0b38b010e771" (UID: "9a483967-d369-48b0-8e8b-0b38b010e771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.115607 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvfnp\" (UniqueName: \"kubernetes.io/projected/9a483967-d369-48b0-8e8b-0b38b010e771-kube-api-access-tvfnp\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.115644 4702 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.115657 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a483967-d369-48b0-8e8b-0b38b010e771-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.497488 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-98k5l" event={"ID":"9a483967-d369-48b0-8e8b-0b38b010e771","Type":"ContainerDied","Data":"cf33e57073ccb91521f88bec1fd582a5c52bf47a210f66345a3caf845cd3e9c9"} Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.497542 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf33e57073ccb91521f88bec1fd582a5c52bf47a210f66345a3caf845cd3e9c9" Oct 07 12:46:29 crc kubenswrapper[4702]: I1007 12:46:29.497575 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-98k5l" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.852250 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 07 12:46:30 crc kubenswrapper[4702]: E1007 12:46:30.852615 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a483967-d369-48b0-8e8b-0b38b010e771" containerName="glance-db-sync" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.852633 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a483967-d369-48b0-8e8b-0b38b010e771" containerName="glance-db-sync" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.852808 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a483967-d369-48b0-8e8b-0b38b010e771" containerName="glance-db-sync" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.854114 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.857064 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.857178 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.857586 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-vv9tn" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.880950 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944477 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944517 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944539 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-dev\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944566 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-logs\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944593 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944612 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944803 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944857 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjrv6\" (UniqueName: \"kubernetes.io/projected/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-kube-api-access-qjrv6\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944886 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.944994 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-run\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.945038 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.945081 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.945145 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-sys\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.945248 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.979890 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.981189 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:30 crc kubenswrapper[4702]: I1007 12:46:30.984127 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.006429 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046411 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046454 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046471 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-dev\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046497 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-logs\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046527 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046542 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046578 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjrv6\" (UniqueName: \"kubernetes.io/projected/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-kube-api-access-qjrv6\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046588 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046605 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-dev\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046598 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046791 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046874 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-run\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046888 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046935 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-run\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046978 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047003 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047004 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047018 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-logs\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047029 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-sys\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047048 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-sys\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.046976 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047014 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047078 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047184 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.047317 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.052592 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.057324 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.061631 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjrv6\" (UniqueName: \"kubernetes.io/projected/2d410bd4-0ed4-4da0-a3cb-8de5825fac11-kube-api-access-qjrv6\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.072956 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.075250 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2d410bd4-0ed4-4da0-a3cb-8de5825fac11\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148489 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148538 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-dev\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148564 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148584 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4t85\" (UniqueName: \"kubernetes.io/projected/1dd9147b-a25d-4082-a79b-9f8f340a7218-kube-api-access-x4t85\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148600 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148613 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-logs\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148635 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148655 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148686 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148706 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148734 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148752 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148811 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.148837 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-sys\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.171031 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250218 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250517 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-logs\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250543 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250563 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250595 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250616 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250648 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250664 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250709 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250723 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-sys\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250744 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250762 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-dev\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250783 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.250801 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4t85\" (UniqueName: \"kubernetes.io/projected/1dd9147b-a25d-4082-a79b-9f8f340a7218-kube-api-access-x4t85\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.251153 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.251701 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-logs\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.251961 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.252134 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.252177 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.252405 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.253003 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.253055 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.253089 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-sys\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.253089 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.253121 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-dev\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.259632 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.260394 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.269094 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4t85\" (UniqueName: \"kubernetes.io/projected/1dd9147b-a25d-4082-a79b-9f8f340a7218-kube-api-access-x4t85\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.272591 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.298698 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.601154 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:31 crc kubenswrapper[4702]: I1007 12:46:31.650619 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.073959 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.499397 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.552826 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerStarted","Data":"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.552870 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerStarted","Data":"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.552882 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerStarted","Data":"8dcc17e23e09cacdee38840a5f059213f501ff099786ad0c14f8b5e410e58424"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.563088 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2d410bd4-0ed4-4da0-a3cb-8de5825fac11","Type":"ContainerStarted","Data":"49f23d29ca949a29708d806094d227de52fe9b6ce7be2c2fc5b3cf7324f380ea"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.563137 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2d410bd4-0ed4-4da0-a3cb-8de5825fac11","Type":"ContainerStarted","Data":"383cbbb36fd815c87c7cc5aef9a91c29cc70c6281caa2cd1eacfbdf961149976"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.563147 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2d410bd4-0ed4-4da0-a3cb-8de5825fac11","Type":"ContainerStarted","Data":"884b5eb14e8e4acbcc1d1e2f15c3e84bfdebcc6a0bf8756030f7439959a4a415"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.563157 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2d410bd4-0ed4-4da0-a3cb-8de5825fac11","Type":"ContainerStarted","Data":"4196ff1f1ebab1e4184a9dd18ca256ac0b621667c45e1696034fe50598ff7991"} Oct 07 12:46:32 crc kubenswrapper[4702]: I1007 12:46:32.658931 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.658907844 podStartE2EDuration="2.658907844s" podCreationTimestamp="2025-10-07 12:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:46:32.645878428 +0000 UTC m=+1348.821614554" watchObservedRunningTime="2025-10-07 12:46:32.658907844 +0000 UTC m=+1348.834643960" Oct 07 12:46:33 crc kubenswrapper[4702]: I1007 12:46:33.572475 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerStarted","Data":"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850"} Oct 07 12:46:33 crc kubenswrapper[4702]: I1007 12:46:33.572765 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-log" containerID="cri-o://7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" gracePeriod=30 Oct 07 12:46:33 crc kubenswrapper[4702]: I1007 12:46:33.572834 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-httpd" containerID="cri-o://3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" gracePeriod=30 Oct 07 12:46:33 crc kubenswrapper[4702]: I1007 12:46:33.572882 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-api" containerID="cri-o://66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" gracePeriod=30 Oct 07 12:46:33 crc kubenswrapper[4702]: I1007 12:46:33.606421 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=4.606402518 podStartE2EDuration="4.606402518s" podCreationTimestamp="2025-10-07 12:46:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:46:33.60440991 +0000 UTC m=+1349.780146026" watchObservedRunningTime="2025-10-07 12:46:33.606402518 +0000 UTC m=+1349.782138634" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.175804 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301410 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-logs\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301487 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-iscsi\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301549 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-var-locks-brick\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301570 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-httpd-run\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301607 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-sys\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301624 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301645 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-config-data\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301639 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301668 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-scripts\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301689 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301707 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-run\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301745 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-run" (OuterVolumeSpecName: "run") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301834 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-dev\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301865 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4t85\" (UniqueName: \"kubernetes.io/projected/1dd9147b-a25d-4082-a79b-9f8f340a7218-kube-api-access-x4t85\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301909 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301935 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-lib-modules\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301940 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-dev" (OuterVolumeSpecName: "dev") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301936 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-logs" (OuterVolumeSpecName: "logs") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301984 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301985 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301961 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-nvme\") pod \"1dd9147b-a25d-4082-a79b-9f8f340a7218\" (UID: \"1dd9147b-a25d-4082-a79b-9f8f340a7218\") " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302129 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.301839 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-sys" (OuterVolumeSpecName: "sys") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302667 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302681 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302691 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-sys\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302701 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-run\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302709 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-dev\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302717 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302725 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302735 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd9147b-a25d-4082-a79b-9f8f340a7218-logs\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.302742 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1dd9147b-a25d-4082-a79b-9f8f340a7218-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.307078 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dd9147b-a25d-4082-a79b-9f8f340a7218-kube-api-access-x4t85" (OuterVolumeSpecName: "kube-api-access-x4t85") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "kube-api-access-x4t85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.308475 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.316018 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.316992 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-scripts" (OuterVolumeSpecName: "scripts") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.372876 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-config-data" (OuterVolumeSpecName: "config-data") pod "1dd9147b-a25d-4082-a79b-9f8f340a7218" (UID: "1dd9147b-a25d-4082-a79b-9f8f340a7218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.404107 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.404146 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.404160 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd9147b-a25d-4082-a79b-9f8f340a7218-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.404174 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4t85\" (UniqueName: \"kubernetes.io/projected/1dd9147b-a25d-4082-a79b-9f8f340a7218-kube-api-access-x4t85\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.404195 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.419837 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.424965 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.505755 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.505790 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.601801 4702 generic.go:334] "Generic (PLEG): container finished" podID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerID="66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" exitCode=143 Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.601848 4702 generic.go:334] "Generic (PLEG): container finished" podID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerID="3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" exitCode=0 Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.601865 4702 generic.go:334] "Generic (PLEG): container finished" podID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerID="7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" exitCode=143 Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.601893 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerDied","Data":"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850"} Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.602082 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerDied","Data":"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c"} Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.602103 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerDied","Data":"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317"} Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.602121 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1dd9147b-a25d-4082-a79b-9f8f340a7218","Type":"ContainerDied","Data":"8dcc17e23e09cacdee38840a5f059213f501ff099786ad0c14f8b5e410e58424"} Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.602142 4702 scope.go:117] "RemoveContainer" containerID="66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.602320 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.658146 4702 scope.go:117] "RemoveContainer" containerID="3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.669083 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.669133 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.685888 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.688875 4702 scope.go:117] "RemoveContainer" containerID="7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" Oct 07 12:46:34 crc kubenswrapper[4702]: E1007 12:46:34.689289 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-api" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.689321 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-api" Oct 07 12:46:34 crc kubenswrapper[4702]: E1007 12:46:34.689355 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-httpd" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.689365 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-httpd" Oct 07 12:46:34 crc kubenswrapper[4702]: E1007 12:46:34.689382 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-log" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.689391 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-log" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.689578 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-httpd" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.689589 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-api" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.689612 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" containerName="glance-log" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.690876 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.695792 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.734174 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.739089 4702 scope.go:117] "RemoveContainer" containerID="66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" Oct 07 12:46:34 crc kubenswrapper[4702]: E1007 12:46:34.739513 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": container with ID starting with 66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850 not found: ID does not exist" containerID="66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.739565 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850"} err="failed to get container status \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": rpc error: code = NotFound desc = could not find container \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": container with ID starting with 66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850 not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.739602 4702 scope.go:117] "RemoveContainer" containerID="3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" Oct 07 12:46:34 crc kubenswrapper[4702]: E1007 12:46:34.739822 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": container with ID starting with 3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c not found: ID does not exist" containerID="3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.739845 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c"} err="failed to get container status \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": rpc error: code = NotFound desc = could not find container \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": container with ID starting with 3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.739860 4702 scope.go:117] "RemoveContainer" containerID="7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" Oct 07 12:46:34 crc kubenswrapper[4702]: E1007 12:46:34.740508 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": container with ID starting with 7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317 not found: ID does not exist" containerID="7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.740531 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317"} err="failed to get container status \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": rpc error: code = NotFound desc = could not find container \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": container with ID starting with 7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317 not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.740544 4702 scope.go:117] "RemoveContainer" containerID="66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.742944 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850"} err="failed to get container status \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": rpc error: code = NotFound desc = could not find container \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": container with ID starting with 66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850 not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.743287 4702 scope.go:117] "RemoveContainer" containerID="3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.746213 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c"} err="failed to get container status \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": rpc error: code = NotFound desc = could not find container \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": container with ID starting with 3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.746261 4702 scope.go:117] "RemoveContainer" containerID="7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.747059 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317"} err="failed to get container status \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": rpc error: code = NotFound desc = could not find container \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": container with ID starting with 7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317 not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.747089 4702 scope.go:117] "RemoveContainer" containerID="66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.747748 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850"} err="failed to get container status \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": rpc error: code = NotFound desc = could not find container \"66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850\": container with ID starting with 66de46ccfae1c650c164ffdad4b6ad630b7efe3f661d90c8292c2078407cf850 not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.747771 4702 scope.go:117] "RemoveContainer" containerID="3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.748079 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c"} err="failed to get container status \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": rpc error: code = NotFound desc = could not find container \"3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c\": container with ID starting with 3a4600b89d4a128e55788148fb37c275fc2e5bfac218a245e3bc2cb991077a3c not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.748113 4702 scope.go:117] "RemoveContainer" containerID="7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.748350 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317"} err="failed to get container status \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": rpc error: code = NotFound desc = could not find container \"7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317\": container with ID starting with 7d8d62b9191656e65c600749340b957187d0f514fd7f345bfe7ac0c8250dc317 not found: ID does not exist" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.809988 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810073 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d6c48a-8f63-4a51-96be-cc41e868a930-logs\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810163 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d6c48a-8f63-4a51-96be-cc41e868a930-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810205 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810258 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d6c48a-8f63-4a51-96be-cc41e868a930-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810302 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810337 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810376 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-run\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810406 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810447 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810477 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8d6c48a-8f63-4a51-96be-cc41e868a930-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810580 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-dev\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810632 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-sys\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.810654 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkj98\" (UniqueName: \"kubernetes.io/projected/f8d6c48a-8f63-4a51-96be-cc41e868a930-kube-api-access-mkj98\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.912304 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.912741 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8d6c48a-8f63-4a51-96be-cc41e868a930-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.912846 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-dev\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.912694 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.912925 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-sys\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913011 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-dev\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913079 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkj98\" (UniqueName: \"kubernetes.io/projected/f8d6c48a-8f63-4a51-96be-cc41e868a930-kube-api-access-mkj98\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913148 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-sys\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913183 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913334 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d6c48a-8f63-4a51-96be-cc41e868a930-logs\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913454 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d6c48a-8f63-4a51-96be-cc41e868a930-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913591 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913724 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d6c48a-8f63-4a51-96be-cc41e868a930-logs\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913231 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913700 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913239 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8d6c48a-8f63-4a51-96be-cc41e868a930-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.913984 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d6c48a-8f63-4a51-96be-cc41e868a930-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914475 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914564 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914627 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914642 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-run\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914693 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914748 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914846 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.914931 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f8d6c48a-8f63-4a51-96be-cc41e868a930-run\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.918080 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d6c48a-8f63-4a51-96be-cc41e868a930-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.919213 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d6c48a-8f63-4a51-96be-cc41e868a930-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.928684 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkj98\" (UniqueName: \"kubernetes.io/projected/f8d6c48a-8f63-4a51-96be-cc41e868a930-kube-api-access-mkj98\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.931510 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:34 crc kubenswrapper[4702]: I1007 12:46:34.949264 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8d6c48a-8f63-4a51-96be-cc41e868a930\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:35 crc kubenswrapper[4702]: I1007 12:46:35.027725 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:35 crc kubenswrapper[4702]: I1007 12:46:35.499040 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 07 12:46:35 crc kubenswrapper[4702]: W1007 12:46:35.514056 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8d6c48a_8f63_4a51_96be_cc41e868a930.slice/crio-bc232da55f75c1df884a3c8f8683b8b1fbc8902d5aff03f967cd644974edd315 WatchSource:0}: Error finding container bc232da55f75c1df884a3c8f8683b8b1fbc8902d5aff03f967cd644974edd315: Status 404 returned error can't find the container with id bc232da55f75c1df884a3c8f8683b8b1fbc8902d5aff03f967cd644974edd315 Oct 07 12:46:35 crc kubenswrapper[4702]: I1007 12:46:35.615913 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f8d6c48a-8f63-4a51-96be-cc41e868a930","Type":"ContainerStarted","Data":"bc232da55f75c1df884a3c8f8683b8b1fbc8902d5aff03f967cd644974edd315"} Oct 07 12:46:36 crc kubenswrapper[4702]: I1007 12:46:36.634133 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f8d6c48a-8f63-4a51-96be-cc41e868a930","Type":"ContainerStarted","Data":"b95f2ac201a28e04ab69b6aa0b6ee98343bae9a8392af4bc833b6e1706c54064"} Oct 07 12:46:36 crc kubenswrapper[4702]: I1007 12:46:36.634564 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f8d6c48a-8f63-4a51-96be-cc41e868a930","Type":"ContainerStarted","Data":"30278fb04ca1a4c386eb78fc6d494d0a44ec2cae6d50395b5af1ce4666fe904d"} Oct 07 12:46:36 crc kubenswrapper[4702]: I1007 12:46:36.634576 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f8d6c48a-8f63-4a51-96be-cc41e868a930","Type":"ContainerStarted","Data":"2568fe1023effcbfe9723807d565cd6411e2e33802288c6ae150423a69ccc41b"} Oct 07 12:46:36 crc kubenswrapper[4702]: I1007 12:46:36.658235 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dd9147b-a25d-4082-a79b-9f8f340a7218" path="/var/lib/kubelet/pods/1dd9147b-a25d-4082-a79b-9f8f340a7218/volumes" Oct 07 12:46:36 crc kubenswrapper[4702]: I1007 12:46:36.667011 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.666992671 podStartE2EDuration="2.666992671s" podCreationTimestamp="2025-10-07 12:46:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:46:36.666987701 +0000 UTC m=+1352.842723827" watchObservedRunningTime="2025-10-07 12:46:36.666992671 +0000 UTC m=+1352.842728787" Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.262993 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.263422 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.263539 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.265167 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8baadc91bcb473e7996507e561b2cb8a6841ec268f00c5e69f87373e5816c4b"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.265364 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://f8baadc91bcb473e7996507e561b2cb8a6841ec268f00c5e69f87373e5816c4b" gracePeriod=600 Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.660009 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="f8baadc91bcb473e7996507e561b2cb8a6841ec268f00c5e69f87373e5816c4b" exitCode=0 Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.660071 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"f8baadc91bcb473e7996507e561b2cb8a6841ec268f00c5e69f87373e5816c4b"} Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.660370 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b"} Oct 07 12:46:39 crc kubenswrapper[4702]: I1007 12:46:39.660391 4702 scope.go:117] "RemoveContainer" containerID="03f130f70c653ad000f84ff22f5df3f9d3bfff20eccc689a18ef25482e6731a7" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.171593 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.173322 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.173466 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.201447 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.206227 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.255039 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.682841 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.682936 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.682958 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.699557 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.702503 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:41 crc kubenswrapper[4702]: I1007 12:46:41.714354 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.029097 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.029736 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.029756 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.072790 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.079319 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.083556 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.722805 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.722874 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.722883 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.743138 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.745327 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:46:45 crc kubenswrapper[4702]: I1007 12:46:45.758662 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.310273 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5txpj"] Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.312182 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.328469 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5txpj"] Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.387860 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phm8b\" (UniqueName: \"kubernetes.io/projected/efcd0620-4a96-4dd8-8e0d-965b0044e19d-kube-api-access-phm8b\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.387951 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-catalog-content\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.387970 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-utilities\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.489238 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phm8b\" (UniqueName: \"kubernetes.io/projected/efcd0620-4a96-4dd8-8e0d-965b0044e19d-kube-api-access-phm8b\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.489314 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-catalog-content\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.489337 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-utilities\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.489843 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-utilities\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.490017 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-catalog-content\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.507440 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phm8b\" (UniqueName: \"kubernetes.io/projected/efcd0620-4a96-4dd8-8e0d-965b0044e19d-kube-api-access-phm8b\") pod \"redhat-marketplace-5txpj\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:45 crc kubenswrapper[4702]: I1007 12:47:45.637608 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:46 crc kubenswrapper[4702]: I1007 12:47:46.079926 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5txpj"] Oct 07 12:47:46 crc kubenswrapper[4702]: I1007 12:47:46.253744 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5txpj" event={"ID":"efcd0620-4a96-4dd8-8e0d-965b0044e19d","Type":"ContainerStarted","Data":"07ae6e7c2db33cc3dad90efb5463e663c9248d2ca14047677c1056001beaa483"} Oct 07 12:47:47 crc kubenswrapper[4702]: I1007 12:47:47.262321 4702 generic.go:334] "Generic (PLEG): container finished" podID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerID="b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d" exitCode=0 Oct 07 12:47:47 crc kubenswrapper[4702]: I1007 12:47:47.262415 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5txpj" event={"ID":"efcd0620-4a96-4dd8-8e0d-965b0044e19d","Type":"ContainerDied","Data":"b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d"} Oct 07 12:47:47 crc kubenswrapper[4702]: I1007 12:47:47.265066 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 12:47:49 crc kubenswrapper[4702]: I1007 12:47:49.909746 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gjlfc"] Oct 07 12:47:49 crc kubenswrapper[4702]: I1007 12:47:49.913857 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:49 crc kubenswrapper[4702]: I1007 12:47:49.924085 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gjlfc"] Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.059243 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6n7b\" (UniqueName: \"kubernetes.io/projected/f0049023-c4ba-4421-8ce8-c78734739a94-kube-api-access-g6n7b\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.059417 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-utilities\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.059469 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-catalog-content\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.161543 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6n7b\" (UniqueName: \"kubernetes.io/projected/f0049023-c4ba-4421-8ce8-c78734739a94-kube-api-access-g6n7b\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.161737 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-utilities\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.161784 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-catalog-content\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.162698 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-catalog-content\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.162716 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-utilities\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.195918 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6n7b\" (UniqueName: \"kubernetes.io/projected/f0049023-c4ba-4421-8ce8-c78734739a94-kube-api-access-g6n7b\") pod \"community-operators-gjlfc\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.240019 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.302418 4702 generic.go:334] "Generic (PLEG): container finished" podID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerID="697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c" exitCode=0 Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.302491 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5txpj" event={"ID":"efcd0620-4a96-4dd8-8e0d-965b0044e19d","Type":"ContainerDied","Data":"697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c"} Oct 07 12:47:50 crc kubenswrapper[4702]: I1007 12:47:50.770198 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gjlfc"] Oct 07 12:47:51 crc kubenswrapper[4702]: I1007 12:47:51.311050 4702 generic.go:334] "Generic (PLEG): container finished" podID="f0049023-c4ba-4421-8ce8-c78734739a94" containerID="67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19" exitCode=0 Oct 07 12:47:51 crc kubenswrapper[4702]: I1007 12:47:51.311139 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjlfc" event={"ID":"f0049023-c4ba-4421-8ce8-c78734739a94","Type":"ContainerDied","Data":"67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19"} Oct 07 12:47:51 crc kubenswrapper[4702]: I1007 12:47:51.311430 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjlfc" event={"ID":"f0049023-c4ba-4421-8ce8-c78734739a94","Type":"ContainerStarted","Data":"5b607c0558573b6ffc7729ca42fd871a2d4b18f9374959d1881ada232bf2bb14"} Oct 07 12:47:51 crc kubenswrapper[4702]: I1007 12:47:51.314838 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5txpj" event={"ID":"efcd0620-4a96-4dd8-8e0d-965b0044e19d","Type":"ContainerStarted","Data":"2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25"} Oct 07 12:47:51 crc kubenswrapper[4702]: I1007 12:47:51.347889 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5txpj" podStartSLOduration=2.66217852 podStartE2EDuration="6.347872474s" podCreationTimestamp="2025-10-07 12:47:45 +0000 UTC" firstStartedPulling="2025-10-07 12:47:47.264820976 +0000 UTC m=+1423.440557092" lastFinishedPulling="2025-10-07 12:47:50.95051493 +0000 UTC m=+1427.126251046" observedRunningTime="2025-10-07 12:47:51.343890139 +0000 UTC m=+1427.519626295" watchObservedRunningTime="2025-10-07 12:47:51.347872474 +0000 UTC m=+1427.523608580" Oct 07 12:47:52 crc kubenswrapper[4702]: I1007 12:47:52.324458 4702 generic.go:334] "Generic (PLEG): container finished" podID="f0049023-c4ba-4421-8ce8-c78734739a94" containerID="aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5" exitCode=0 Oct 07 12:47:52 crc kubenswrapper[4702]: I1007 12:47:52.324531 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjlfc" event={"ID":"f0049023-c4ba-4421-8ce8-c78734739a94","Type":"ContainerDied","Data":"aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5"} Oct 07 12:47:53 crc kubenswrapper[4702]: I1007 12:47:53.334410 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjlfc" event={"ID":"f0049023-c4ba-4421-8ce8-c78734739a94","Type":"ContainerStarted","Data":"513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3"} Oct 07 12:47:53 crc kubenswrapper[4702]: I1007 12:47:53.360608 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gjlfc" podStartSLOduration=2.607847935 podStartE2EDuration="4.360588127s" podCreationTimestamp="2025-10-07 12:47:49 +0000 UTC" firstStartedPulling="2025-10-07 12:47:51.314250085 +0000 UTC m=+1427.489986201" lastFinishedPulling="2025-10-07 12:47:53.066990267 +0000 UTC m=+1429.242726393" observedRunningTime="2025-10-07 12:47:53.351589458 +0000 UTC m=+1429.527325594" watchObservedRunningTime="2025-10-07 12:47:53.360588127 +0000 UTC m=+1429.536324253" Oct 07 12:47:55 crc kubenswrapper[4702]: I1007 12:47:55.638178 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:55 crc kubenswrapper[4702]: I1007 12:47:55.638648 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:55 crc kubenswrapper[4702]: I1007 12:47:55.723837 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:56 crc kubenswrapper[4702]: I1007 12:47:56.410094 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:56 crc kubenswrapper[4702]: I1007 12:47:56.878578 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5txpj"] Oct 07 12:47:58 crc kubenswrapper[4702]: I1007 12:47:58.391018 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5txpj" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="registry-server" containerID="cri-o://2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25" gracePeriod=2 Oct 07 12:47:58 crc kubenswrapper[4702]: I1007 12:47:58.889414 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:58 crc kubenswrapper[4702]: I1007 12:47:58.998553 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-catalog-content\") pod \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " Oct 07 12:47:58 crc kubenswrapper[4702]: I1007 12:47:58.998692 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phm8b\" (UniqueName: \"kubernetes.io/projected/efcd0620-4a96-4dd8-8e0d-965b0044e19d-kube-api-access-phm8b\") pod \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " Oct 07 12:47:58 crc kubenswrapper[4702]: I1007 12:47:58.999625 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-utilities\") pod \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\" (UID: \"efcd0620-4a96-4dd8-8e0d-965b0044e19d\") " Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.000365 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-utilities" (OuterVolumeSpecName: "utilities") pod "efcd0620-4a96-4dd8-8e0d-965b0044e19d" (UID: "efcd0620-4a96-4dd8-8e0d-965b0044e19d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.008978 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efcd0620-4a96-4dd8-8e0d-965b0044e19d-kube-api-access-phm8b" (OuterVolumeSpecName: "kube-api-access-phm8b") pod "efcd0620-4a96-4dd8-8e0d-965b0044e19d" (UID: "efcd0620-4a96-4dd8-8e0d-965b0044e19d"). InnerVolumeSpecName "kube-api-access-phm8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.014471 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efcd0620-4a96-4dd8-8e0d-965b0044e19d" (UID: "efcd0620-4a96-4dd8-8e0d-965b0044e19d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.101735 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.101785 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phm8b\" (UniqueName: \"kubernetes.io/projected/efcd0620-4a96-4dd8-8e0d-965b0044e19d-kube-api-access-phm8b\") on node \"crc\" DevicePath \"\"" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.101805 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efcd0620-4a96-4dd8-8e0d-965b0044e19d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.404760 4702 generic.go:334] "Generic (PLEG): container finished" podID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerID="2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25" exitCode=0 Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.405059 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5txpj" event={"ID":"efcd0620-4a96-4dd8-8e0d-965b0044e19d","Type":"ContainerDied","Data":"2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25"} Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.405154 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5txpj" event={"ID":"efcd0620-4a96-4dd8-8e0d-965b0044e19d","Type":"ContainerDied","Data":"07ae6e7c2db33cc3dad90efb5463e663c9248d2ca14047677c1056001beaa483"} Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.405197 4702 scope.go:117] "RemoveContainer" containerID="2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.405510 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5txpj" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.444087 4702 scope.go:117] "RemoveContainer" containerID="697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.454997 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5txpj"] Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.466162 4702 scope.go:117] "RemoveContainer" containerID="b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.483730 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5txpj"] Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.501235 4702 scope.go:117] "RemoveContainer" containerID="2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25" Oct 07 12:47:59 crc kubenswrapper[4702]: E1007 12:47:59.501763 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25\": container with ID starting with 2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25 not found: ID does not exist" containerID="2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.501810 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25"} err="failed to get container status \"2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25\": rpc error: code = NotFound desc = could not find container \"2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25\": container with ID starting with 2309aacae027d34f271b6c044a2fb496f5aaa1f9071a8efaddec636ad070da25 not found: ID does not exist" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.501843 4702 scope.go:117] "RemoveContainer" containerID="697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c" Oct 07 12:47:59 crc kubenswrapper[4702]: E1007 12:47:59.502258 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c\": container with ID starting with 697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c not found: ID does not exist" containerID="697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.502294 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c"} err="failed to get container status \"697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c\": rpc error: code = NotFound desc = could not find container \"697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c\": container with ID starting with 697955f00e999e996ae6aa0687bd8d30c127d9e80bfad052886156bf06c5b86c not found: ID does not exist" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.502320 4702 scope.go:117] "RemoveContainer" containerID="b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d" Oct 07 12:47:59 crc kubenswrapper[4702]: E1007 12:47:59.502832 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d\": container with ID starting with b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d not found: ID does not exist" containerID="b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d" Oct 07 12:47:59 crc kubenswrapper[4702]: I1007 12:47:59.502947 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d"} err="failed to get container status \"b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d\": rpc error: code = NotFound desc = could not find container \"b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d\": container with ID starting with b008897d1a9d0b517dfaad9df9974409d1c955fcd18009048fc904136124f87d not found: ID does not exist" Oct 07 12:48:00 crc kubenswrapper[4702]: I1007 12:48:00.240546 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:48:00 crc kubenswrapper[4702]: I1007 12:48:00.240646 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:48:00 crc kubenswrapper[4702]: I1007 12:48:00.322961 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:48:00 crc kubenswrapper[4702]: I1007 12:48:00.496561 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:48:00 crc kubenswrapper[4702]: I1007 12:48:00.656324 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" path="/var/lib/kubelet/pods/efcd0620-4a96-4dd8-8e0d-965b0044e19d/volumes" Oct 07 12:48:02 crc kubenswrapper[4702]: I1007 12:48:02.285401 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gjlfc"] Oct 07 12:48:02 crc kubenswrapper[4702]: I1007 12:48:02.436566 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gjlfc" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="registry-server" containerID="cri-o://513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3" gracePeriod=2 Oct 07 12:48:02 crc kubenswrapper[4702]: I1007 12:48:02.919112 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.003496 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6n7b\" (UniqueName: \"kubernetes.io/projected/f0049023-c4ba-4421-8ce8-c78734739a94-kube-api-access-g6n7b\") pod \"f0049023-c4ba-4421-8ce8-c78734739a94\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.003581 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-catalog-content\") pod \"f0049023-c4ba-4421-8ce8-c78734739a94\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.011407 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0049023-c4ba-4421-8ce8-c78734739a94-kube-api-access-g6n7b" (OuterVolumeSpecName: "kube-api-access-g6n7b") pod "f0049023-c4ba-4421-8ce8-c78734739a94" (UID: "f0049023-c4ba-4421-8ce8-c78734739a94"). InnerVolumeSpecName "kube-api-access-g6n7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.071847 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0049023-c4ba-4421-8ce8-c78734739a94" (UID: "f0049023-c4ba-4421-8ce8-c78734739a94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.104948 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-utilities\") pod \"f0049023-c4ba-4421-8ce8-c78734739a94\" (UID: \"f0049023-c4ba-4421-8ce8-c78734739a94\") " Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.105466 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6n7b\" (UniqueName: \"kubernetes.io/projected/f0049023-c4ba-4421-8ce8-c78734739a94-kube-api-access-g6n7b\") on node \"crc\" DevicePath \"\"" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.105491 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.106071 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-utilities" (OuterVolumeSpecName: "utilities") pod "f0049023-c4ba-4421-8ce8-c78734739a94" (UID: "f0049023-c4ba-4421-8ce8-c78734739a94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.206873 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0049023-c4ba-4421-8ce8-c78734739a94-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.446597 4702 generic.go:334] "Generic (PLEG): container finished" podID="f0049023-c4ba-4421-8ce8-c78734739a94" containerID="513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3" exitCode=0 Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.446673 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjlfc" event={"ID":"f0049023-c4ba-4421-8ce8-c78734739a94","Type":"ContainerDied","Data":"513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3"} Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.446732 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjlfc" event={"ID":"f0049023-c4ba-4421-8ce8-c78734739a94","Type":"ContainerDied","Data":"5b607c0558573b6ffc7729ca42fd871a2d4b18f9374959d1881ada232bf2bb14"} Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.446766 4702 scope.go:117] "RemoveContainer" containerID="513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.446995 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjlfc" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.482366 4702 scope.go:117] "RemoveContainer" containerID="aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.500479 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gjlfc"] Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.512407 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gjlfc"] Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.516585 4702 scope.go:117] "RemoveContainer" containerID="67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.552336 4702 scope.go:117] "RemoveContainer" containerID="513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3" Oct 07 12:48:03 crc kubenswrapper[4702]: E1007 12:48:03.552813 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3\": container with ID starting with 513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3 not found: ID does not exist" containerID="513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.552865 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3"} err="failed to get container status \"513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3\": rpc error: code = NotFound desc = could not find container \"513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3\": container with ID starting with 513592c39cd408111221548ddd10e53695759a8d45fb512f537098d74bc43bd3 not found: ID does not exist" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.552956 4702 scope.go:117] "RemoveContainer" containerID="aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5" Oct 07 12:48:03 crc kubenswrapper[4702]: E1007 12:48:03.553444 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5\": container with ID starting with aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5 not found: ID does not exist" containerID="aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.553508 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5"} err="failed to get container status \"aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5\": rpc error: code = NotFound desc = could not find container \"aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5\": container with ID starting with aadb8c6a4d4aae387f6cbd8bc58633488e5ee3e875bca273c512db495986e0c5 not found: ID does not exist" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.553546 4702 scope.go:117] "RemoveContainer" containerID="67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19" Oct 07 12:48:03 crc kubenswrapper[4702]: E1007 12:48:03.554138 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19\": container with ID starting with 67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19 not found: ID does not exist" containerID="67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19" Oct 07 12:48:03 crc kubenswrapper[4702]: I1007 12:48:03.554183 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19"} err="failed to get container status \"67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19\": rpc error: code = NotFound desc = could not find container \"67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19\": container with ID starting with 67806d763162927d4c6e66783fe580e85adf49888fd3a9c4021d99ee8293bd19 not found: ID does not exist" Oct 07 12:48:04 crc kubenswrapper[4702]: I1007 12:48:04.653685 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" path="/var/lib/kubelet/pods/f0049023-c4ba-4421-8ce8-c78734739a94/volumes" Oct 07 12:48:39 crc kubenswrapper[4702]: I1007 12:48:39.263108 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:48:39 crc kubenswrapper[4702]: I1007 12:48:39.265174 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.891631 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m577t"] Oct 07 12:49:03 crc kubenswrapper[4702]: E1007 12:49:03.892753 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="extract-content" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.892768 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="extract-content" Oct 07 12:49:03 crc kubenswrapper[4702]: E1007 12:49:03.892820 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="registry-server" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.892832 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="registry-server" Oct 07 12:49:03 crc kubenswrapper[4702]: E1007 12:49:03.892885 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="extract-utilities" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.892919 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="extract-utilities" Oct 07 12:49:03 crc kubenswrapper[4702]: E1007 12:49:03.892931 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="extract-utilities" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.892937 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="extract-utilities" Oct 07 12:49:03 crc kubenswrapper[4702]: E1007 12:49:03.892948 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="extract-content" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.892954 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="extract-content" Oct 07 12:49:03 crc kubenswrapper[4702]: E1007 12:49:03.892976 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="registry-server" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.893040 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="registry-server" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.893210 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="efcd0620-4a96-4dd8-8e0d-965b0044e19d" containerName="registry-server" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.893224 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0049023-c4ba-4421-8ce8-c78734739a94" containerName="registry-server" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.895692 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:03 crc kubenswrapper[4702]: I1007 12:49:03.931568 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m577t"] Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.032113 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-utilities\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.032165 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vghnl\" (UniqueName: \"kubernetes.io/projected/64c3de44-32c9-4861-8417-d8de29e4525b-kube-api-access-vghnl\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.032358 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-catalog-content\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.134067 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-utilities\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.134125 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vghnl\" (UniqueName: \"kubernetes.io/projected/64c3de44-32c9-4861-8417-d8de29e4525b-kube-api-access-vghnl\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.134200 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-catalog-content\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.134618 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-utilities\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.134693 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-catalog-content\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.157175 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vghnl\" (UniqueName: \"kubernetes.io/projected/64c3de44-32c9-4861-8417-d8de29e4525b-kube-api-access-vghnl\") pod \"redhat-operators-m577t\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.231517 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:04 crc kubenswrapper[4702]: I1007 12:49:04.698869 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m577t"] Oct 07 12:49:05 crc kubenswrapper[4702]: I1007 12:49:05.017274 4702 generic.go:334] "Generic (PLEG): container finished" podID="64c3de44-32c9-4861-8417-d8de29e4525b" containerID="1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc" exitCode=0 Oct 07 12:49:05 crc kubenswrapper[4702]: I1007 12:49:05.017381 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m577t" event={"ID":"64c3de44-32c9-4861-8417-d8de29e4525b","Type":"ContainerDied","Data":"1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc"} Oct 07 12:49:05 crc kubenswrapper[4702]: I1007 12:49:05.017529 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m577t" event={"ID":"64c3de44-32c9-4861-8417-d8de29e4525b","Type":"ContainerStarted","Data":"a06b83b30d4a1611e562cd6281d8b8e3f2a27ee38f0765598e5922e52ced3425"} Oct 07 12:49:07 crc kubenswrapper[4702]: I1007 12:49:07.038150 4702 generic.go:334] "Generic (PLEG): container finished" podID="64c3de44-32c9-4861-8417-d8de29e4525b" containerID="9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a" exitCode=0 Oct 07 12:49:07 crc kubenswrapper[4702]: I1007 12:49:07.038211 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m577t" event={"ID":"64c3de44-32c9-4861-8417-d8de29e4525b","Type":"ContainerDied","Data":"9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a"} Oct 07 12:49:08 crc kubenswrapper[4702]: I1007 12:49:08.048242 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m577t" event={"ID":"64c3de44-32c9-4861-8417-d8de29e4525b","Type":"ContainerStarted","Data":"c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f"} Oct 07 12:49:09 crc kubenswrapper[4702]: I1007 12:49:09.262736 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:49:09 crc kubenswrapper[4702]: I1007 12:49:09.263843 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:49:14 crc kubenswrapper[4702]: I1007 12:49:14.232408 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:14 crc kubenswrapper[4702]: I1007 12:49:14.233063 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:14 crc kubenswrapper[4702]: I1007 12:49:14.304880 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:14 crc kubenswrapper[4702]: I1007 12:49:14.339475 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m577t" podStartSLOduration=8.68712148 podStartE2EDuration="11.339446373s" podCreationTimestamp="2025-10-07 12:49:03 +0000 UTC" firstStartedPulling="2025-10-07 12:49:05.018621587 +0000 UTC m=+1501.194357703" lastFinishedPulling="2025-10-07 12:49:07.67094647 +0000 UTC m=+1503.846682596" observedRunningTime="2025-10-07 12:49:08.065689104 +0000 UTC m=+1504.241425220" watchObservedRunningTime="2025-10-07 12:49:14.339446373 +0000 UTC m=+1510.515182529" Oct 07 12:49:15 crc kubenswrapper[4702]: I1007 12:49:15.190783 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:15 crc kubenswrapper[4702]: I1007 12:49:15.253342 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m577t"] Oct 07 12:49:17 crc kubenswrapper[4702]: I1007 12:49:17.142445 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m577t" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="registry-server" containerID="cri-o://c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f" gracePeriod=2 Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.715910 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.807871 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-utilities\") pod \"64c3de44-32c9-4861-8417-d8de29e4525b\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.807992 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vghnl\" (UniqueName: \"kubernetes.io/projected/64c3de44-32c9-4861-8417-d8de29e4525b-kube-api-access-vghnl\") pod \"64c3de44-32c9-4861-8417-d8de29e4525b\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.808021 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-catalog-content\") pod \"64c3de44-32c9-4861-8417-d8de29e4525b\" (UID: \"64c3de44-32c9-4861-8417-d8de29e4525b\") " Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.809773 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-utilities" (OuterVolumeSpecName: "utilities") pod "64c3de44-32c9-4861-8417-d8de29e4525b" (UID: "64c3de44-32c9-4861-8417-d8de29e4525b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.815539 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c3de44-32c9-4861-8417-d8de29e4525b-kube-api-access-vghnl" (OuterVolumeSpecName: "kube-api-access-vghnl") pod "64c3de44-32c9-4861-8417-d8de29e4525b" (UID: "64c3de44-32c9-4861-8417-d8de29e4525b"). InnerVolumeSpecName "kube-api-access-vghnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.897684 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64c3de44-32c9-4861-8417-d8de29e4525b" (UID: "64c3de44-32c9-4861-8417-d8de29e4525b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.910133 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.910170 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vghnl\" (UniqueName: \"kubernetes.io/projected/64c3de44-32c9-4861-8417-d8de29e4525b-kube-api-access-vghnl\") on node \"crc\" DevicePath \"\"" Oct 07 12:49:18 crc kubenswrapper[4702]: I1007 12:49:18.910187 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c3de44-32c9-4861-8417-d8de29e4525b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.164999 4702 generic.go:334] "Generic (PLEG): container finished" podID="64c3de44-32c9-4861-8417-d8de29e4525b" containerID="c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f" exitCode=0 Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.165061 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m577t" event={"ID":"64c3de44-32c9-4861-8417-d8de29e4525b","Type":"ContainerDied","Data":"c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f"} Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.165103 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m577t" event={"ID":"64c3de44-32c9-4861-8417-d8de29e4525b","Type":"ContainerDied","Data":"a06b83b30d4a1611e562cd6281d8b8e3f2a27ee38f0765598e5922e52ced3425"} Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.165132 4702 scope.go:117] "RemoveContainer" containerID="c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.165238 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m577t" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.197608 4702 scope.go:117] "RemoveContainer" containerID="9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.221433 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m577t"] Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.234360 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m577t"] Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.250845 4702 scope.go:117] "RemoveContainer" containerID="1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.283182 4702 scope.go:117] "RemoveContainer" containerID="c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f" Oct 07 12:49:19 crc kubenswrapper[4702]: E1007 12:49:19.283710 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f\": container with ID starting with c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f not found: ID does not exist" containerID="c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.283761 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f"} err="failed to get container status \"c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f\": rpc error: code = NotFound desc = could not find container \"c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f\": container with ID starting with c2fd715499c935b060e50cd39f6a9e2c50835e155d492f618af325e7a8d3506f not found: ID does not exist" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.283787 4702 scope.go:117] "RemoveContainer" containerID="9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a" Oct 07 12:49:19 crc kubenswrapper[4702]: E1007 12:49:19.284122 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a\": container with ID starting with 9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a not found: ID does not exist" containerID="9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.284156 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a"} err="failed to get container status \"9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a\": rpc error: code = NotFound desc = could not find container \"9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a\": container with ID starting with 9cd1b3d26765f2a8b9fa93754ace089ae010cc11272dfe5915ed58c127e6938a not found: ID does not exist" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.284174 4702 scope.go:117] "RemoveContainer" containerID="1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc" Oct 07 12:49:19 crc kubenswrapper[4702]: E1007 12:49:19.284455 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc\": container with ID starting with 1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc not found: ID does not exist" containerID="1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc" Oct 07 12:49:19 crc kubenswrapper[4702]: I1007 12:49:19.284494 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc"} err="failed to get container status \"1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc\": rpc error: code = NotFound desc = could not find container \"1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc\": container with ID starting with 1ff64e9744bb55150f35350a912d85d82cc12e6ea0cdb753ad4768b421b23fcc not found: ID does not exist" Oct 07 12:49:20 crc kubenswrapper[4702]: I1007 12:49:20.657737 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" path="/var/lib/kubelet/pods/64c3de44-32c9-4861-8417-d8de29e4525b/volumes" Oct 07 12:49:39 crc kubenswrapper[4702]: I1007 12:49:39.263118 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:49:39 crc kubenswrapper[4702]: I1007 12:49:39.263949 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:49:39 crc kubenswrapper[4702]: I1007 12:49:39.264015 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" Oct 07 12:49:39 crc kubenswrapper[4702]: I1007 12:49:39.264821 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b"} pod="openshift-machine-config-operator/machine-config-daemon-w4btn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 12:49:39 crc kubenswrapper[4702]: I1007 12:49:39.264947 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" containerID="cri-o://bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" gracePeriod=600 Oct 07 12:49:39 crc kubenswrapper[4702]: E1007 12:49:39.402015 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:49:40 crc kubenswrapper[4702]: I1007 12:49:40.372799 4702 generic.go:334] "Generic (PLEG): container finished" podID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" exitCode=0 Oct 07 12:49:40 crc kubenswrapper[4702]: I1007 12:49:40.372868 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerDied","Data":"bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b"} Oct 07 12:49:40 crc kubenswrapper[4702]: I1007 12:49:40.372946 4702 scope.go:117] "RemoveContainer" containerID="f8baadc91bcb473e7996507e561b2cb8a6841ec268f00c5e69f87373e5816c4b" Oct 07 12:49:40 crc kubenswrapper[4702]: I1007 12:49:40.373324 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:49:40 crc kubenswrapper[4702]: E1007 12:49:40.373646 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:49:53 crc kubenswrapper[4702]: I1007 12:49:53.641789 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:49:53 crc kubenswrapper[4702]: E1007 12:49:53.643554 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:50:08 crc kubenswrapper[4702]: I1007 12:50:08.642477 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:50:08 crc kubenswrapper[4702]: E1007 12:50:08.643462 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:50:14 crc kubenswrapper[4702]: I1007 12:50:14.338690 4702 scope.go:117] "RemoveContainer" containerID="d432120e5e9ff79f44484441786c8c50c2520dd944f2a88a69032bbd41379e4f" Oct 07 12:50:14 crc kubenswrapper[4702]: I1007 12:50:14.404982 4702 scope.go:117] "RemoveContainer" containerID="5d4efdb54fbe55f8540e6cddc57ce58e31d4873d8d4320bef170963ce5793520" Oct 07 12:50:14 crc kubenswrapper[4702]: I1007 12:50:14.430953 4702 scope.go:117] "RemoveContainer" containerID="d7a73528c3e779b416f4869f51c2d9dd3aadd82148a8b69059676a2837bfe6e3" Oct 07 12:50:20 crc kubenswrapper[4702]: I1007 12:50:20.640938 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:50:20 crc kubenswrapper[4702]: E1007 12:50:20.641727 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:50:35 crc kubenswrapper[4702]: I1007 12:50:35.641453 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:50:35 crc kubenswrapper[4702]: E1007 12:50:35.643518 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:50:50 crc kubenswrapper[4702]: I1007 12:50:50.641850 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:50:50 crc kubenswrapper[4702]: E1007 12:50:50.642958 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:50:53 crc kubenswrapper[4702]: I1007 12:50:53.058380 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-p6frh"] Oct 07 12:50:53 crc kubenswrapper[4702]: I1007 12:50:53.063531 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-p6frh"] Oct 07 12:50:54 crc kubenswrapper[4702]: I1007 12:50:54.653177 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a" path="/var/lib/kubelet/pods/29ad0c3f-7b1b-454d-aeb0-fe6f52f96c7a/volumes" Oct 07 12:51:03 crc kubenswrapper[4702]: I1007 12:51:03.032881 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-cccd-account-create-wb8v7"] Oct 07 12:51:03 crc kubenswrapper[4702]: I1007 12:51:03.040014 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-cccd-account-create-wb8v7"] Oct 07 12:51:03 crc kubenswrapper[4702]: I1007 12:51:03.641530 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:51:03 crc kubenswrapper[4702]: E1007 12:51:03.641999 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:51:04 crc kubenswrapper[4702]: I1007 12:51:04.661251 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80c8e674-d14d-4594-b2c6-a6d3e92ab9f1" path="/var/lib/kubelet/pods/80c8e674-d14d-4594-b2c6-a6d3e92ab9f1/volumes" Oct 07 12:51:14 crc kubenswrapper[4702]: I1007 12:51:14.531498 4702 scope.go:117] "RemoveContainer" containerID="8024eddcc1b69fd6713bf0fe38024280c2862eb43138e4ca3019a4aed51f3c89" Oct 07 12:51:14 crc kubenswrapper[4702]: I1007 12:51:14.565047 4702 scope.go:117] "RemoveContainer" containerID="eef78e54a2813c17fc204ba6096e8e15ebfa7991450f73a04ba75388f26c4652" Oct 07 12:51:18 crc kubenswrapper[4702]: I1007 12:51:18.641077 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:51:18 crc kubenswrapper[4702]: E1007 12:51:18.641805 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:51:31 crc kubenswrapper[4702]: I1007 12:51:31.640317 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:51:31 crc kubenswrapper[4702]: E1007 12:51:31.641135 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:51:42 crc kubenswrapper[4702]: I1007 12:51:42.640794 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:51:42 crc kubenswrapper[4702]: E1007 12:51:42.641829 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:51:53 crc kubenswrapper[4702]: I1007 12:51:53.641024 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:51:53 crc kubenswrapper[4702]: E1007 12:51:53.642045 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:52:05 crc kubenswrapper[4702]: I1007 12:52:05.642261 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:52:05 crc kubenswrapper[4702]: E1007 12:52:05.642837 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:52:14 crc kubenswrapper[4702]: I1007 12:52:14.694896 4702 scope.go:117] "RemoveContainer" containerID="ad83c31644dda87324a229c2c270095d7acd9251bea3421771a4291a343a5aaf" Oct 07 12:52:14 crc kubenswrapper[4702]: I1007 12:52:14.753508 4702 scope.go:117] "RemoveContainer" containerID="9436437b6624d68ba66378b440c9ee2e84990c45aff60b742a9d8f2001566ac9" Oct 07 12:52:14 crc kubenswrapper[4702]: I1007 12:52:14.777038 4702 scope.go:117] "RemoveContainer" containerID="815d9b7862ceaecf7d22ffd5809c84dfb3724c0633100dea2ce54e64903feb50" Oct 07 12:52:16 crc kubenswrapper[4702]: I1007 12:52:16.039723 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-8dlbs"] Oct 07 12:52:16 crc kubenswrapper[4702]: I1007 12:52:16.046998 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-8dlbs"] Oct 07 12:52:16 crc kubenswrapper[4702]: I1007 12:52:16.650749 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e6332da-21f3-4136-94e9-c5302ba865ac" path="/var/lib/kubelet/pods/3e6332da-21f3-4136-94e9-c5302ba865ac/volumes" Oct 07 12:52:19 crc kubenswrapper[4702]: I1007 12:52:19.640471 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:52:19 crc kubenswrapper[4702]: E1007 12:52:19.641067 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:52:23 crc kubenswrapper[4702]: I1007 12:52:23.027706 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-fxd5x"] Oct 07 12:52:23 crc kubenswrapper[4702]: I1007 12:52:23.034358 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-fxd5x"] Oct 07 12:52:24 crc kubenswrapper[4702]: I1007 12:52:24.660368 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="749ab4c9-2ed8-4e0d-b905-bd8e6719d322" path="/var/lib/kubelet/pods/749ab4c9-2ed8-4e0d-b905-bd8e6719d322/volumes" Oct 07 12:52:31 crc kubenswrapper[4702]: I1007 12:52:31.641453 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:52:31 crc kubenswrapper[4702]: E1007 12:52:31.642523 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:52:44 crc kubenswrapper[4702]: I1007 12:52:44.650545 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:52:44 crc kubenswrapper[4702]: E1007 12:52:44.652655 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.487961 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:52:52 crc kubenswrapper[4702]: E1007 12:52:52.489016 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="extract-content" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.489040 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="extract-content" Oct 07 12:52:52 crc kubenswrapper[4702]: E1007 12:52:52.489074 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="extract-utilities" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.489084 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="extract-utilities" Oct 07 12:52:52 crc kubenswrapper[4702]: E1007 12:52:52.489119 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="registry-server" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.489131 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="registry-server" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.489415 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c3de44-32c9-4861-8417-d8de29e4525b" containerName="registry-server" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.490104 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.492206 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-kf8kh" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.492484 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.492960 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.493742 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.525006 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.577079 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-config\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.577159 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l59w\" (UniqueName: \"kubernetes.io/projected/6ee54c5f-69b4-4b16-9501-77871f522845-kube-api-access-2l59w\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.577238 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-scripts\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.577261 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-config-secret\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.684625 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-scripts\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.684691 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-config-secret\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.684762 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-config\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.684838 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l59w\" (UniqueName: \"kubernetes.io/projected/6ee54c5f-69b4-4b16-9501-77871f522845-kube-api-access-2l59w\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.686000 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-scripts\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.686193 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-config\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.702579 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6ee54c5f-69b4-4b16-9501-77871f522845-openstack-config-secret\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.705071 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l59w\" (UniqueName: \"kubernetes.io/projected/6ee54c5f-69b4-4b16-9501-77871f522845-kube-api-access-2l59w\") pod \"openstackclient\" (UID: \"6ee54c5f-69b4-4b16-9501-77871f522845\") " pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:52 crc kubenswrapper[4702]: I1007 12:52:52.826664 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 07 12:52:53 crc kubenswrapper[4702]: I1007 12:52:53.070671 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 07 12:52:53 crc kubenswrapper[4702]: I1007 12:52:53.145849 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"6ee54c5f-69b4-4b16-9501-77871f522845","Type":"ContainerStarted","Data":"cac4ec5a55b7cb05d3bdd2dc4f6eeb4de01e3cc443c4bbe55c28354bbfdd63f3"} Oct 07 12:52:54 crc kubenswrapper[4702]: I1007 12:52:54.158047 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"6ee54c5f-69b4-4b16-9501-77871f522845","Type":"ContainerStarted","Data":"7dfb6658c152208421abf1db08a71408a827cc3b39bdd02a462a6f7ba0cacc80"} Oct 07 12:52:54 crc kubenswrapper[4702]: I1007 12:52:54.180051 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.180023785 podStartE2EDuration="2.180023785s" podCreationTimestamp="2025-10-07 12:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 12:52:54.179183011 +0000 UTC m=+1730.354919187" watchObservedRunningTime="2025-10-07 12:52:54.180023785 +0000 UTC m=+1730.355759931" Oct 07 12:52:56 crc kubenswrapper[4702]: I1007 12:52:56.641250 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:52:56 crc kubenswrapper[4702]: E1007 12:52:56.641772 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:53:09 crc kubenswrapper[4702]: I1007 12:53:09.641666 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:53:09 crc kubenswrapper[4702]: E1007 12:53:09.642616 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:53:14 crc kubenswrapper[4702]: I1007 12:53:14.887364 4702 scope.go:117] "RemoveContainer" containerID="d2485110569ea9a25f6e03d3cfe7eab686bbd47024adaadfe40eab7e09526720" Oct 07 12:53:14 crc kubenswrapper[4702]: I1007 12:53:14.931817 4702 scope.go:117] "RemoveContainer" containerID="629e04090a084a22efea3b46523dec96fd0e1c8b526acf9c768142b67ce59c5a" Oct 07 12:53:21 crc kubenswrapper[4702]: I1007 12:53:21.654529 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:53:21 crc kubenswrapper[4702]: E1007 12:53:21.655266 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:53:36 crc kubenswrapper[4702]: I1007 12:53:36.641198 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:53:36 crc kubenswrapper[4702]: E1007 12:53:36.642355 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:53:48 crc kubenswrapper[4702]: I1007 12:53:48.641406 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:53:48 crc kubenswrapper[4702]: E1007 12:53:48.642657 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:54:03 crc kubenswrapper[4702]: I1007 12:54:03.641480 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:54:03 crc kubenswrapper[4702]: E1007 12:54:03.642453 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.611445 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m8rqd/must-gather-t6t29"] Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.613184 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.616676 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m8rqd"/"openshift-service-ca.crt" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.617280 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m8rqd"/"kube-root-ca.crt" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.630989 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m8rqd/must-gather-t6t29"] Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.632533 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmwq4\" (UniqueName: \"kubernetes.io/projected/3830702d-c8fa-4b5d-8352-e47f15c9afed-kube-api-access-xmwq4\") pod \"must-gather-t6t29\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.632685 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3830702d-c8fa-4b5d-8352-e47f15c9afed-must-gather-output\") pod \"must-gather-t6t29\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.734081 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmwq4\" (UniqueName: \"kubernetes.io/projected/3830702d-c8fa-4b5d-8352-e47f15c9afed-kube-api-access-xmwq4\") pod \"must-gather-t6t29\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.734191 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3830702d-c8fa-4b5d-8352-e47f15c9afed-must-gather-output\") pod \"must-gather-t6t29\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.734585 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3830702d-c8fa-4b5d-8352-e47f15c9afed-must-gather-output\") pod \"must-gather-t6t29\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.752465 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmwq4\" (UniqueName: \"kubernetes.io/projected/3830702d-c8fa-4b5d-8352-e47f15c9afed-kube-api-access-xmwq4\") pod \"must-gather-t6t29\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:07 crc kubenswrapper[4702]: I1007 12:54:07.935242 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:54:08 crc kubenswrapper[4702]: I1007 12:54:08.171978 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m8rqd/must-gather-t6t29"] Oct 07 12:54:08 crc kubenswrapper[4702]: I1007 12:54:08.181956 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 12:54:08 crc kubenswrapper[4702]: I1007 12:54:08.839618 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8rqd/must-gather-t6t29" event={"ID":"3830702d-c8fa-4b5d-8352-e47f15c9afed","Type":"ContainerStarted","Data":"94d6e57f82a75d0738ba981049f548a00e5bc164be0511dcb80fab0046ed7fe3"} Oct 07 12:54:13 crc kubenswrapper[4702]: I1007 12:54:13.897475 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8rqd/must-gather-t6t29" event={"ID":"3830702d-c8fa-4b5d-8352-e47f15c9afed","Type":"ContainerStarted","Data":"b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd"} Oct 07 12:54:13 crc kubenswrapper[4702]: I1007 12:54:13.898056 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8rqd/must-gather-t6t29" event={"ID":"3830702d-c8fa-4b5d-8352-e47f15c9afed","Type":"ContainerStarted","Data":"aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c"} Oct 07 12:54:13 crc kubenswrapper[4702]: I1007 12:54:13.912624 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m8rqd/must-gather-t6t29" podStartSLOduration=2.186276067 podStartE2EDuration="6.912609257s" podCreationTimestamp="2025-10-07 12:54:07 +0000 UTC" firstStartedPulling="2025-10-07 12:54:08.181871507 +0000 UTC m=+1804.357607633" lastFinishedPulling="2025-10-07 12:54:12.908204707 +0000 UTC m=+1809.083940823" observedRunningTime="2025-10-07 12:54:13.909919849 +0000 UTC m=+1810.085655995" watchObservedRunningTime="2025-10-07 12:54:13.912609257 +0000 UTC m=+1810.088345373" Oct 07 12:54:15 crc kubenswrapper[4702]: I1007 12:54:15.640498 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:54:15 crc kubenswrapper[4702]: E1007 12:54:15.641862 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.694606 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d4h9c"] Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.697881 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.712932 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4h9c"] Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.872244 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-catalog-content\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.872299 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgpj4\" (UniqueName: \"kubernetes.io/projected/ca98b3b8-baff-4fae-a20a-95946a7c38cf-kube-api-access-zgpj4\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.872323 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-utilities\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.973702 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-catalog-content\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.973778 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgpj4\" (UniqueName: \"kubernetes.io/projected/ca98b3b8-baff-4fae-a20a-95946a7c38cf-kube-api-access-zgpj4\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.973805 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-utilities\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.974318 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-utilities\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.974654 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-catalog-content\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:23 crc kubenswrapper[4702]: I1007 12:54:23.991308 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgpj4\" (UniqueName: \"kubernetes.io/projected/ca98b3b8-baff-4fae-a20a-95946a7c38cf-kube-api-access-zgpj4\") pod \"certified-operators-d4h9c\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:24 crc kubenswrapper[4702]: I1007 12:54:24.027268 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:24 crc kubenswrapper[4702]: I1007 12:54:24.541515 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4h9c"] Oct 07 12:54:24 crc kubenswrapper[4702]: I1007 12:54:24.970800 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerStarted","Data":"71d060a2fa358743d1fa8187d95ffe05e1c81ab34357500b5237e27a0b074e41"} Oct 07 12:54:26 crc kubenswrapper[4702]: I1007 12:54:26.988630 4702 generic.go:334] "Generic (PLEG): container finished" podID="ca98b3b8-baff-4fae-a20a-95946a7c38cf" containerID="0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738" exitCode=0 Oct 07 12:54:26 crc kubenswrapper[4702]: I1007 12:54:26.988720 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerDied","Data":"0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738"} Oct 07 12:54:29 crc kubenswrapper[4702]: I1007 12:54:29.004256 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerStarted","Data":"187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6"} Oct 07 12:54:30 crc kubenswrapper[4702]: I1007 12:54:30.014118 4702 generic.go:334] "Generic (PLEG): container finished" podID="ca98b3b8-baff-4fae-a20a-95946a7c38cf" containerID="187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6" exitCode=0 Oct 07 12:54:30 crc kubenswrapper[4702]: I1007 12:54:30.014217 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerDied","Data":"187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6"} Oct 07 12:54:30 crc kubenswrapper[4702]: I1007 12:54:30.641127 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:54:30 crc kubenswrapper[4702]: E1007 12:54:30.641524 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w4btn_openshift-machine-config-operator(e1ca41d2-75d4-4871-abe5-46a3e172fadd)\"" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" Oct 07 12:54:31 crc kubenswrapper[4702]: I1007 12:54:31.024730 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerStarted","Data":"fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27"} Oct 07 12:54:31 crc kubenswrapper[4702]: I1007 12:54:31.052981 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d4h9c" podStartSLOduration=4.479961702 podStartE2EDuration="8.052960656s" podCreationTimestamp="2025-10-07 12:54:23 +0000 UTC" firstStartedPulling="2025-10-07 12:54:26.990311102 +0000 UTC m=+1823.166047218" lastFinishedPulling="2025-10-07 12:54:30.563310036 +0000 UTC m=+1826.739046172" observedRunningTime="2025-10-07 12:54:31.052667417 +0000 UTC m=+1827.228403533" watchObservedRunningTime="2025-10-07 12:54:31.052960656 +0000 UTC m=+1827.228696772" Oct 07 12:54:34 crc kubenswrapper[4702]: I1007 12:54:34.028018 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:34 crc kubenswrapper[4702]: I1007 12:54:34.028400 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:34 crc kubenswrapper[4702]: I1007 12:54:34.068485 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:44 crc kubenswrapper[4702]: I1007 12:54:44.126288 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:44 crc kubenswrapper[4702]: I1007 12:54:44.648593 4702 scope.go:117] "RemoveContainer" containerID="bb0b0e06a9afbeca2cfbb02b2f0a22b02447012360db5b155725e95179d3ee2b" Oct 07 12:54:45 crc kubenswrapper[4702]: I1007 12:54:45.139474 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" event={"ID":"e1ca41d2-75d4-4871-abe5-46a3e172fadd","Type":"ContainerStarted","Data":"c55e44836f54c79e29d3c107813197d7c432733bbef4a28099b5f8fa27f4b51c"} Oct 07 12:54:47 crc kubenswrapper[4702]: I1007 12:54:47.682065 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4h9c"] Oct 07 12:54:47 crc kubenswrapper[4702]: I1007 12:54:47.682998 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d4h9c" podUID="ca98b3b8-baff-4fae-a20a-95946a7c38cf" containerName="registry-server" containerID="cri-o://fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27" gracePeriod=2 Oct 07 12:54:47 crc kubenswrapper[4702]: I1007 12:54:47.849727 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/util/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.038201 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/util/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.094546 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/pull/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.129357 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/pull/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.132924 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.179800 4702 generic.go:334] "Generic (PLEG): container finished" podID="ca98b3b8-baff-4fae-a20a-95946a7c38cf" containerID="fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27" exitCode=0 Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.179835 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerDied","Data":"fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27"} Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.179858 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4h9c" event={"ID":"ca98b3b8-baff-4fae-a20a-95946a7c38cf","Type":"ContainerDied","Data":"71d060a2fa358743d1fa8187d95ffe05e1c81ab34357500b5237e27a0b074e41"} Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.179876 4702 scope.go:117] "RemoveContainer" containerID="fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.180008 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4h9c" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.195970 4702 scope.go:117] "RemoveContainer" containerID="187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.226285 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgpj4\" (UniqueName: \"kubernetes.io/projected/ca98b3b8-baff-4fae-a20a-95946a7c38cf-kube-api-access-zgpj4\") pod \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.226393 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-catalog-content\") pod \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.226415 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-utilities\") pod \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\" (UID: \"ca98b3b8-baff-4fae-a20a-95946a7c38cf\") " Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.227363 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-utilities" (OuterVolumeSpecName: "utilities") pod "ca98b3b8-baff-4fae-a20a-95946a7c38cf" (UID: "ca98b3b8-baff-4fae-a20a-95946a7c38cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.234097 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca98b3b8-baff-4fae-a20a-95946a7c38cf-kube-api-access-zgpj4" (OuterVolumeSpecName: "kube-api-access-zgpj4") pod "ca98b3b8-baff-4fae-a20a-95946a7c38cf" (UID: "ca98b3b8-baff-4fae-a20a-95946a7c38cf"). InnerVolumeSpecName "kube-api-access-zgpj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.240434 4702 scope.go:117] "RemoveContainer" containerID="0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.275525 4702 scope.go:117] "RemoveContainer" containerID="fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27" Oct 07 12:54:48 crc kubenswrapper[4702]: E1007 12:54:48.275873 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27\": container with ID starting with fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27 not found: ID does not exist" containerID="fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.275930 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27"} err="failed to get container status \"fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27\": rpc error: code = NotFound desc = could not find container \"fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27\": container with ID starting with fadf230c43dbf82db6ec3b8ca757eb1dead48cbe5c37a3314b6ec1b3de046e27 not found: ID does not exist" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.275958 4702 scope.go:117] "RemoveContainer" containerID="187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6" Oct 07 12:54:48 crc kubenswrapper[4702]: E1007 12:54:48.276258 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6\": container with ID starting with 187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6 not found: ID does not exist" containerID="187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.276302 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6"} err="failed to get container status \"187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6\": rpc error: code = NotFound desc = could not find container \"187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6\": container with ID starting with 187d798eda467a9ffe2c3772bc9cacc42a5561d05e11ccc2e9afbd779d5d01c6 not found: ID does not exist" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.276332 4702 scope.go:117] "RemoveContainer" containerID="0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738" Oct 07 12:54:48 crc kubenswrapper[4702]: E1007 12:54:48.276565 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738\": container with ID starting with 0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738 not found: ID does not exist" containerID="0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.276585 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738"} err="failed to get container status \"0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738\": rpc error: code = NotFound desc = could not find container \"0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738\": container with ID starting with 0a1d48245cfc32f1361ada5db85d59320086c5c1b1ae2a0f9c3dfc363ee90738 not found: ID does not exist" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.277821 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca98b3b8-baff-4fae-a20a-95946a7c38cf" (UID: "ca98b3b8-baff-4fae-a20a-95946a7c38cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.285418 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/pull/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.313052 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/util/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.327783 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.327813 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca98b3b8-baff-4fae-a20a-95946a7c38cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.327826 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgpj4\" (UniqueName: \"kubernetes.io/projected/ca98b3b8-baff-4fae-a20a-95946a7c38cf-kube-api-access-zgpj4\") on node \"crc\" DevicePath \"\"" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.358621 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10f57ec60de4df58ed39de93369cc80174e5ad08476bc9cf01944ad890r727t_9b6ccce5-1320-48c5-b65f-6f0cfd8d771e/extract/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.434113 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/util/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.505328 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4h9c"] Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.511760 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d4h9c"] Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.642524 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/util/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.648925 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca98b3b8-baff-4fae-a20a-95946a7c38cf" path="/var/lib/kubelet/pods/ca98b3b8-baff-4fae-a20a-95946a7c38cf/volumes" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.652616 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/pull/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.698933 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/pull/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.827644 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/util/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.827741 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/extract/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.869025 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1b0eb0f74eb5f610f434f2930f528c582e44b95e8d8fb6ec09a20c02fap84l4_ac694397-13ed-4daa-9a05-7ae908c3da4f/pull/0.log" Oct 07 12:54:48 crc kubenswrapper[4702]: I1007 12:54:48.990338 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/util/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.170920 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/pull/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.181181 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/util/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.186505 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/pull/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.305758 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/util/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.371240 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/pull/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.371509 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb402f945d54a80688ab4565a4e9d19d5d4eb730a5ce0fdf7f49eb3138g4t2_cdfc6f48-f0c8-4c8e-97f0-2df10778f9f8/extract/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.485191 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/util/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.680856 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/pull/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.696292 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/util/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.718076 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/pull/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.859392 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/util/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.880106 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/pull/0.log" Oct 07 12:54:49 crc kubenswrapper[4702]: I1007 12:54:49.884500 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590z2tsh_1bf04750-2468-4c4d-8931-b03f97a6b36a/extract/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.029881 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/util/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.171665 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/pull/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.205255 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/pull/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.217505 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/util/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.371537 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/util/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.372817 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/pull/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.520233 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a010b1fa02645b98b2846ddaf7b57b0b5fb7c211c004326d6c1d71a32efmjp4_96e269e9-1b43-40d4-8590-ee9290396d92/extract/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.573942 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/util/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.701216 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/util/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.727260 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/pull/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.727369 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/pull/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.921835 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/util/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.922139 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/extract/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.928027 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ee3e80c0d89273140d050d464665e565aa58ef37fb933a47da0d9153c8xl6fp_e819cdf1-7dc2-436f-9d41-c9f4c2783957/pull/0.log" Oct 07 12:54:50 crc kubenswrapper[4702]: I1007 12:54:50.955589 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/util/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.124546 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/pull/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.128960 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/pull/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.144243 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/util/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.304483 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/pull/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.310239 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/util/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.344911 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f2628378dfbe9d43c4c77358844c5bb7d39b0ec6a549d0614459ab45bebwfnf_d73554d4-7d8e-4f1b-a49f-866642b8822c/extract/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.371687 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-85f8db96c5-w6ds7_fac7e760-a6e5-42dd-8cdd-ebcf25bba335/kube-rbac-proxy/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.510592 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-lg4mj_63310df2-8f1e-4ddc-b41f-f338b754812e/registry-server/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.551800 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f9cd5475c-4k5dc_bcc3148d-161a-46ea-910c-aec0e3b3c824/kube-rbac-proxy/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.555858 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-85f8db96c5-w6ds7_fac7e760-a6e5-42dd-8cdd-ebcf25bba335/manager/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.665935 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f9cd5475c-4k5dc_bcc3148d-161a-46ea-910c-aec0e3b3c824/manager/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.732009 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-snplx_1032492a-6e8e-4795-ada8-1a3f84943f07/registry-server/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.797042 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7dc9c6fbcb-wzs5p_beb78a59-5cba-42ca-a7bb-5351e77835b8/kube-rbac-proxy/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.878590 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7dc9c6fbcb-wzs5p_beb78a59-5cba-42ca-a7bb-5351e77835b8/manager/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.966810 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-2zrr4_da437e15-80b4-4b13-8c1e-a117c6084274/registry-server/0.log" Oct 07 12:54:51 crc kubenswrapper[4702]: I1007 12:54:51.970984 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-556f5598f7-x5kzf_444fbf26-f791-432e-8051-2bb3995dcd6f/kube-rbac-proxy/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.081878 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-556f5598f7-x5kzf_444fbf26-f791-432e-8051-2bb3995dcd6f/manager/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.124111 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c58767fd-n8qwp_126878fb-9a6e-4cf2-9bd5-cfcfab6888c6/kube-rbac-proxy/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.170627 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-r6rq2_1a57bec4-d2af-4a75-b137-e3bce61d7aef/registry-server/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.281119 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c58767fd-n8qwp_126878fb-9a6e-4cf2-9bd5-cfcfab6888c6/manager/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.345207 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-xn9xd_0591047e-35dd-4650-8730-d7795ce53a56/registry-server/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.404818 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-zr2gn_feb903e7-e1ea-4519-9913-f68dfb45c9e1/operator/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.524262 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-pphjk_ec9bb282-6aca-4099-901d-b5bc2c780f43/registry-server/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.570250 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6c7959756-6fx5l_1a4a43d0-4ae6-4ccf-927a-014abbb8be62/kube-rbac-proxy/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.592752 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6c7959756-6fx5l_1a4a43d0-4ae6-4ccf-927a-014abbb8be62/manager/0.log" Oct 07 12:54:52 crc kubenswrapper[4702]: I1007 12:54:52.702471 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-j8lzr_68a99f85-bca1-4976-a914-7f83cfb77ac4/registry-server/0.log" Oct 07 12:55:07 crc kubenswrapper[4702]: I1007 12:55:07.172137 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-47q4n_802177f9-2e2e-4cb8-895f-686ed017c9e3/control-plane-machine-set-operator/0.log" Oct 07 12:55:07 crc kubenswrapper[4702]: I1007 12:55:07.326594 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qtvlc_5abee0e6-194a-4457-86dc-6d27a745589b/machine-api-operator/0.log" Oct 07 12:55:07 crc kubenswrapper[4702]: I1007 12:55:07.348142 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qtvlc_5abee0e6-194a-4457-86dc-6d27a745589b/kube-rbac-proxy/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.460929 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-bwfk4_1d618973-0b94-496b-b7eb-d06b37ec6506/kube-rbac-proxy/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.534549 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-bwfk4_1d618973-0b94-496b-b7eb-d06b37ec6506/controller/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.643121 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-frr-files/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.854949 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-frr-files/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.865500 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-reloader/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.873526 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-reloader/0.log" Oct 07 12:55:23 crc kubenswrapper[4702]: I1007 12:55:23.882089 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-metrics/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.069741 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-metrics/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.076058 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-frr-files/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.090854 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-reloader/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.102887 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-metrics/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.237829 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-reloader/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.242104 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-metrics/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.264400 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/cp-frr-files/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.309431 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/controller/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.410289 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/frr-metrics/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.432461 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/kube-rbac-proxy/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.502148 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/kube-rbac-proxy-frr/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.589539 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/reloader/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.684766 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-8w262_26b44d11-5310-42a7-bbbc-fc95de8743a2/frr-k8s-webhook-server/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.824667 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6nskv_3f9fc314-be24-41d6-80bd-e4622cd1804e/frr/0.log" Oct 07 12:55:24 crc kubenswrapper[4702]: I1007 12:55:24.893971 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6fdccd459c-wjxtf_10a60231-0c5a-463a-8836-310df99ffe9d/manager/0.log" Oct 07 12:55:25 crc kubenswrapper[4702]: I1007 12:55:25.028906 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5966c949c-9xjc8_190a6066-4fa2-4dbe-bc52-6125fad41859/webhook-server/0.log" Oct 07 12:55:25 crc kubenswrapper[4702]: I1007 12:55:25.080339 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7vt7t_234a63eb-cada-437c-9df9-d2b1058ba5d4/kube-rbac-proxy/0.log" Oct 07 12:55:25 crc kubenswrapper[4702]: I1007 12:55:25.203081 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7vt7t_234a63eb-cada-437c-9df9-d2b1058ba5d4/speaker/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.112093 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-7db2-account-create-mkb7h_1c75a1f9-755f-4b5e-9e45-88fb60ce3596/mariadb-account-create/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.224166 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-6kvrs_598ee324-eed2-48a2-b64d-2e439a2002af/mariadb-database-create/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.269235 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-98k5l_9a483967-d369-48b0-8e8b-0b38b010e771/glance-db-sync/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.400682 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_2d410bd4-0ed4-4da0-a3cb-8de5825fac11/glance-api/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.409815 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_2d410bd4-0ed4-4da0-a3cb-8de5825fac11/glance-httpd/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.458932 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_2d410bd4-0ed4-4da0-a3cb-8de5825fac11/glance-log/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.565087 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_f8d6c48a-8f63-4a51-96be-cc41e868a930/glance-api/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.575239 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_f8d6c48a-8f63-4a51-96be-cc41e868a930/glance-httpd/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.619590 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_f8d6c48a-8f63-4a51-96be-cc41e868a930/glance-log/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.926991 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_6cc5554b-8ac6-4893-bc94-3cfc4f815438/memcached/0.log" Oct 07 12:55:38 crc kubenswrapper[4702]: I1007 12:55:38.987685 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_f1b04b98-3506-4aff-afd5-f9a2730576f6/mysql-bootstrap/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.041318 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-69787db686-hgjdk_95471c20-4000-4f65-9f81-255c1b909f35/keystone-api/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.213424 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_f1b04b98-3506-4aff-afd5-f9a2730576f6/mysql-bootstrap/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.220620 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_f1b04b98-3506-4aff-afd5-f9a2730576f6/galera/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.316338 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_94961522-a5fc-43a6-a62b-2bba41a15be9/mysql-bootstrap/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.472647 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_94961522-a5fc-43a6-a62b-2bba41a15be9/mysql-bootstrap/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.491537 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_94961522-a5fc-43a6-a62b-2bba41a15be9/galera/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.519045 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_f1ba8472-9f12-45f8-9dc5-562c2b936b18/mysql-bootstrap/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.692739 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_f1ba8472-9f12-45f8-9dc5-562c2b936b18/galera/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.706070 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_6ee54c5f-69b4-4b16-9501-77871f522845/openstackclient/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.730377 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_f1ba8472-9f12-45f8-9dc5-562c2b936b18/mysql-bootstrap/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.878359 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_c3cf606b-7113-4066-8808-ebde118e2247/setup-container/0.log" Oct 07 12:55:39 crc kubenswrapper[4702]: I1007 12:55:39.994587 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_c3cf606b-7113-4066-8808-ebde118e2247/rabbitmq/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.015699 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_c3cf606b-7113-4066-8808-ebde118e2247/setup-container/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.085502 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6dd8f59749-9f2cn_0b851329-4872-417e-84c5-24b567021d51/proxy-httpd/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.194843 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6dd8f59749-9f2cn_0b851329-4872-417e-84c5-24b567021d51/proxy-server/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.220746 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-gv9w7_5d1fc69f-1999-40a0-acf6-eba215843669/swift-ring-rebalance/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.364051 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/account-reaper/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.387765 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/account-auditor/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.391821 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/account-replicator/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.426550 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/account-server/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.542314 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/container-auditor/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.553357 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/container-server/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.559594 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/container-replicator/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.599005 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/container-updater/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.700820 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/object-expirer/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.701315 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/object-auditor/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.726269 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/object-replicator/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.777752 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/object-server/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.857317 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/object-updater/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.864838 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/rsync/0.log" Oct 07 12:55:40 crc kubenswrapper[4702]: I1007 12:55:40.910791 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7646503c-ea03-4ce3-9438-54b187e1aec3/swift-recon-cron/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.403303 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/util/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.521537 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/util/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.541253 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/pull/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.553032 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/pull/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.704547 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/pull/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.727168 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/extract/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.765182 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2stb6j_d93c1009-09dc-4db3-b49a-6981a02756be/util/0.log" Oct 07 12:55:53 crc kubenswrapper[4702]: I1007 12:55:53.964373 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/extract-utilities/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.106006 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/extract-utilities/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.114412 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/extract-content/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.125964 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/extract-content/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.304606 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/extract-utilities/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.304771 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/extract-content/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.513917 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/extract-utilities/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.719919 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/extract-content/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.720114 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/extract-content/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.735276 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-86dwm_8b3119f5-1f29-4f75-ad92-ace559a6b148/registry-server/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.796731 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/extract-utilities/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.889659 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/extract-utilities/0.log" Oct 07 12:55:54 crc kubenswrapper[4702]: I1007 12:55:54.908132 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/extract-content/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.115534 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8r8v4_76e55658-e2c9-4e68-b04a-a8ebd620e355/marketplace-operator/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.211792 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/extract-utilities/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.317272 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2ntvb_de46fdcb-23b3-4c75-baf4-584a8b29d3d3/registry-server/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.359265 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/extract-utilities/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.385709 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/extract-content/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.409414 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/extract-content/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.578758 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/extract-content/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.590774 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/extract-utilities/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.696266 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m4xtt_bc501a5c-2f7a-4668-a581-6d8b2d0191e8/registry-server/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.771955 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/extract-utilities/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.943775 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/extract-utilities/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.946046 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/extract-content/0.log" Oct 07 12:55:55 crc kubenswrapper[4702]: I1007 12:55:55.954729 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/extract-content/0.log" Oct 07 12:55:56 crc kubenswrapper[4702]: I1007 12:55:56.079546 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/extract-utilities/0.log" Oct 07 12:55:56 crc kubenswrapper[4702]: I1007 12:55:56.136315 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/extract-content/0.log" Oct 07 12:55:56 crc kubenswrapper[4702]: I1007 12:55:56.390955 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-krkwq_49f95513-76af-43cd-ab53-5be28017d39d/registry-server/0.log" Oct 07 12:56:11 crc kubenswrapper[4702]: I1007 12:56:11.036304 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-6kvrs"] Oct 07 12:56:11 crc kubenswrapper[4702]: I1007 12:56:11.043168 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-6kvrs"] Oct 07 12:56:12 crc kubenswrapper[4702]: I1007 12:56:12.674623 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="598ee324-eed2-48a2-b64d-2e439a2002af" path="/var/lib/kubelet/pods/598ee324-eed2-48a2-b64d-2e439a2002af/volumes" Oct 07 12:56:15 crc kubenswrapper[4702]: I1007 12:56:15.054126 4702 scope.go:117] "RemoveContainer" containerID="9997cce2b6960ff81d9e8af96d39546a9d7c65142dea8b839e47443813e725c0" Oct 07 12:56:20 crc kubenswrapper[4702]: I1007 12:56:20.029294 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-7db2-account-create-mkb7h"] Oct 07 12:56:20 crc kubenswrapper[4702]: I1007 12:56:20.036092 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-7db2-account-create-mkb7h"] Oct 07 12:56:20 crc kubenswrapper[4702]: I1007 12:56:20.654790 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c75a1f9-755f-4b5e-9e45-88fb60ce3596" path="/var/lib/kubelet/pods/1c75a1f9-755f-4b5e-9e45-88fb60ce3596/volumes" Oct 07 12:56:29 crc kubenswrapper[4702]: I1007 12:56:29.045274 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-98k5l"] Oct 07 12:56:29 crc kubenswrapper[4702]: I1007 12:56:29.049613 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-98k5l"] Oct 07 12:56:30 crc kubenswrapper[4702]: I1007 12:56:30.648507 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a483967-d369-48b0-8e8b-0b38b010e771" path="/var/lib/kubelet/pods/9a483967-d369-48b0-8e8b-0b38b010e771/volumes" Oct 07 12:56:59 crc kubenswrapper[4702]: I1007 12:56:59.298348 4702 generic.go:334] "Generic (PLEG): container finished" podID="3830702d-c8fa-4b5d-8352-e47f15c9afed" containerID="aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c" exitCode=0 Oct 07 12:56:59 crc kubenswrapper[4702]: I1007 12:56:59.298785 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8rqd/must-gather-t6t29" event={"ID":"3830702d-c8fa-4b5d-8352-e47f15c9afed","Type":"ContainerDied","Data":"aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c"} Oct 07 12:56:59 crc kubenswrapper[4702]: I1007 12:56:59.299337 4702 scope.go:117] "RemoveContainer" containerID="aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c" Oct 07 12:56:59 crc kubenswrapper[4702]: I1007 12:56:59.456428 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8rqd_must-gather-t6t29_3830702d-c8fa-4b5d-8352-e47f15c9afed/gather/0.log" Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.347303 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m8rqd/must-gather-t6t29"] Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.347967 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m8rqd/must-gather-t6t29" podUID="3830702d-c8fa-4b5d-8352-e47f15c9afed" containerName="copy" containerID="cri-o://b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd" gracePeriod=2 Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.360249 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m8rqd/must-gather-t6t29"] Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.763766 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8rqd_must-gather-t6t29_3830702d-c8fa-4b5d-8352-e47f15c9afed/copy/0.log" Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.767528 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.880648 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmwq4\" (UniqueName: \"kubernetes.io/projected/3830702d-c8fa-4b5d-8352-e47f15c9afed-kube-api-access-xmwq4\") pod \"3830702d-c8fa-4b5d-8352-e47f15c9afed\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.880801 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3830702d-c8fa-4b5d-8352-e47f15c9afed-must-gather-output\") pod \"3830702d-c8fa-4b5d-8352-e47f15c9afed\" (UID: \"3830702d-c8fa-4b5d-8352-e47f15c9afed\") " Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.895952 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3830702d-c8fa-4b5d-8352-e47f15c9afed-kube-api-access-xmwq4" (OuterVolumeSpecName: "kube-api-access-xmwq4") pod "3830702d-c8fa-4b5d-8352-e47f15c9afed" (UID: "3830702d-c8fa-4b5d-8352-e47f15c9afed"). InnerVolumeSpecName "kube-api-access-xmwq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.964854 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3830702d-c8fa-4b5d-8352-e47f15c9afed-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3830702d-c8fa-4b5d-8352-e47f15c9afed" (UID: "3830702d-c8fa-4b5d-8352-e47f15c9afed"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.982397 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmwq4\" (UniqueName: \"kubernetes.io/projected/3830702d-c8fa-4b5d-8352-e47f15c9afed-kube-api-access-xmwq4\") on node \"crc\" DevicePath \"\"" Oct 07 12:57:06 crc kubenswrapper[4702]: I1007 12:57:06.982435 4702 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3830702d-c8fa-4b5d-8352-e47f15c9afed-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.366281 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8rqd_must-gather-t6t29_3830702d-c8fa-4b5d-8352-e47f15c9afed/copy/0.log" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.366686 4702 generic.go:334] "Generic (PLEG): container finished" podID="3830702d-c8fa-4b5d-8352-e47f15c9afed" containerID="b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd" exitCode=143 Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.366752 4702 scope.go:117] "RemoveContainer" containerID="b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.366762 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8rqd/must-gather-t6t29" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.388824 4702 scope.go:117] "RemoveContainer" containerID="aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.445041 4702 scope.go:117] "RemoveContainer" containerID="b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd" Oct 07 12:57:07 crc kubenswrapper[4702]: E1007 12:57:07.445712 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd\": container with ID starting with b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd not found: ID does not exist" containerID="b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.445743 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd"} err="failed to get container status \"b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd\": rpc error: code = NotFound desc = could not find container \"b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd\": container with ID starting with b523404f06fcfa1bc701939505b29c3ac333978f5ab39e385e355b9c5fb172bd not found: ID does not exist" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.445769 4702 scope.go:117] "RemoveContainer" containerID="aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c" Oct 07 12:57:07 crc kubenswrapper[4702]: E1007 12:57:07.446081 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c\": container with ID starting with aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c not found: ID does not exist" containerID="aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c" Oct 07 12:57:07 crc kubenswrapper[4702]: I1007 12:57:07.446103 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c"} err="failed to get container status \"aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c\": rpc error: code = NotFound desc = could not find container \"aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c\": container with ID starting with aa45cb12c67c1ff9406c502b43e7067c9679b1f424779c9cf98bcc146241039c not found: ID does not exist" Oct 07 12:57:08 crc kubenswrapper[4702]: I1007 12:57:08.651041 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3830702d-c8fa-4b5d-8352-e47f15c9afed" path="/var/lib/kubelet/pods/3830702d-c8fa-4b5d-8352-e47f15c9afed/volumes" Oct 07 12:57:09 crc kubenswrapper[4702]: I1007 12:57:09.263306 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:57:09 crc kubenswrapper[4702]: I1007 12:57:09.263365 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 12:57:15 crc kubenswrapper[4702]: I1007 12:57:15.114069 4702 scope.go:117] "RemoveContainer" containerID="b2afe95112d7ef6a6054011ec90bf2d3e9fab48538ce3e2d1ee5ee5792ac256c" Oct 07 12:57:15 crc kubenswrapper[4702]: I1007 12:57:15.173271 4702 scope.go:117] "RemoveContainer" containerID="61a1d71a3dfc9f1d2b96dd869abc1fc2a28836d26b627c02cf4eecdc3cc6c263" Oct 07 12:57:39 crc kubenswrapper[4702]: I1007 12:57:39.263395 4702 patch_prober.go:28] interesting pod/machine-config-daemon-w4btn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 12:57:39 crc kubenswrapper[4702]: I1007 12:57:39.264112 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w4btn" podUID="e1ca41d2-75d4-4871-abe5-46a3e172fadd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071207130024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071207131017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071202710016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071202710015450 5ustar corecore